You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2018/06/20 17:23:42 UTC
[JENKINS] Lucene-Solr-7.x-Linux (32bit/jdk1.8.0_172) - Build # 2162
- Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2162/
Java: 32bit/jdk1.8.0_172 -server -XX:+UseParallelGC
1 tests failed.
FAILED: org.apache.solr.cloud.MoveReplicaHDFSTest.testFailedMove
Error Message:
No live SolrServers available to handle this request:[https://127.0.0.1:46543/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:46571/solr/MoveReplicaHDFSTest_failed_coll_true]
Stack Trace:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request:[https://127.0.0.1:46543/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:46571/solr/MoveReplicaHDFSTest_failed_coll_true]
at __randomizedtesting.SeedInfo.seed([6AEEF24E2527C591:C02321BC92F41041]:0)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:462)
at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1106)
at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:886)
at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:993)
at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:819)
at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:942)
at org.apache.solr.cloud.MoveReplicaTest.testFailedMove(MoveReplicaTest.java:288)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at https://127.0.0.1:46543/solr/MoveReplicaHDFSTest_failed_coll_true: no servers hosting shard: shard1
at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:436)
... 46 more
Build Log:
[...truncated 14490 lines...]
[junit4] Suite: org.apache.solr.cloud.MoveReplicaHDFSTest
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/init-core-data-001
[junit4] 2> 1557820 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 4 servers in /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001
[junit4] 2> 1557820 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1557820 INFO (Thread-2741) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1557820 INFO (Thread-2741) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 1557822 ERROR (Thread-2741) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 1557920 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.s.c.ZkTestServer start zk server on port:41683
[junit4] 2> 1557922 INFO (zkConnectionManagerCallback-3071-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1557927 INFO (jetty-launcher-3068-thread-4) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 1557927 INFO (jetty-launcher-3068-thread-2) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 1557927 INFO (jetty-launcher-3068-thread-3) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 1557927 INFO (jetty-launcher-3068-thread-1) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-4) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-4) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-4) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@16a1d25{/solr,null,AVAILABLE}
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-4) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@1cf150f{/solr,null,AVAILABLE}
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.e.j.s.AbstractConnector Started ServerConnector@1bbc15b{SSL,[ssl, http/1.1]}{127.0.0.1:34477}
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.e.j.s.Server Started @1557959ms
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-4) [ ] o.e.j.s.AbstractConnector Started ServerConnector@8d4d57{SSL,[ssl, http/1.1]}{127.0.0.1:46571}
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=34477}
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-4) [ ] o.e.j.s.Server Started @1557959ms
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=46571}
[junit4] 2> 1557928 ERROR (jetty-launcher-3068-thread-3) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1557928 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1557929 ERROR (jetty-launcher-3068-thread-4) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T16:48:56.481Z
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T16:48:56.481Z
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-2) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-2) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1557929 INFO (jetty-launcher-3068-thread-2) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1557929 INFO (zkConnectionManagerCallback-3073-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-1) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-2) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@1b9f92b{/solr,null,AVAILABLE}
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@1fed583{/solr,null,AVAILABLE}
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-2) [ ] o.e.j.s.AbstractConnector Started ServerConnector@b5a870{SSL,[ssl, http/1.1]}{127.0.0.1:40189}
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@57ed22{SSL,[ssl, http/1.1]}{127.0.0.1:40599}
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-2) [ ] o.e.j.s.Server Started @1557961ms
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-1) [ ] o.e.j.s.Server Started @1557961ms
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=40189}
[junit4] 2> 1557930 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=40599}
[junit4] 2> 1557931 ERROR (jetty-launcher-3068-thread-2) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1557931 ERROR (jetty-launcher-3068-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T16:48:56.483Z
[junit4] 2> 1557931 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T16:48:56.483Z
[junit4] 2> 1557935 INFO (zkConnectionManagerCallback-3075-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1557936 INFO (zkConnectionManagerCallback-3079-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1557936 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1557936 INFO (zkConnectionManagerCallback-3078-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1557936 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1557939 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1558263 INFO (jetty-launcher-3068-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41683/solr
[junit4] 2> 1558283 INFO (zkConnectionManagerCallback-3083-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558296 INFO (zkConnectionManagerCallback-3085-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558327 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1558327 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:40599_solr
[junit4] 2> 1558327 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.c.Overseer Overseer (id=72234177385791494-127.0.0.1:40599_solr-n_0000000000) starting
[junit4] 2> 1558372 INFO (zkConnectionManagerCallback-3092-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558373 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41683/solr ready
[junit4] 2> 1558396 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:40599_solr
[junit4] 2> 1558412 INFO (zkCallback-3084-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1558415 INFO (zkCallback-3091-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1558427 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1558447 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40599.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558454 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40599.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558454 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40599.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558455 INFO (jetty-launcher-3068-thread-1) [n:127.0.0.1:40599_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node1/.
[junit4] 2> 1558666 INFO (jetty-launcher-3068-thread-2) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41683/solr
[junit4] 2> 1558667 INFO (zkConnectionManagerCallback-3097-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558675 INFO (zkConnectionManagerCallback-3099-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558678 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1558679 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1558680 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1558680 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:40189_solr
[junit4] 2> 1558681 INFO (zkCallback-3084-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1558681 INFO (zkCallback-3091-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1558681 INFO (zkCallback-3098-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1558697 INFO (zkConnectionManagerCallback-3106-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558698 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1558698 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41683/solr ready
[junit4] 2> 1558698 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1558711 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40189.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558717 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40189.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558717 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40189.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558718 INFO (jetty-launcher-3068-thread-2) [n:127.0.0.1:40189_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node2/.
[junit4] 2> 1558875 INFO (jetty-launcher-3068-thread-3) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41683/solr
[junit4] 2> 1558875 INFO (zkConnectionManagerCallback-3111-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558877 INFO (zkConnectionManagerCallback-3113-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558879 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1558880 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1558881 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1558881 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34477_solr
[junit4] 2> 1558881 INFO (zkCallback-3091-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1558881 INFO (zkCallback-3098-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1558882 INFO (zkCallback-3105-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1558883 INFO (zkCallback-3084-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1558883 INFO (zkCallback-3112-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1558896 INFO (zkConnectionManagerCallback-3120-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1558896 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 1558897 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41683/solr ready
[junit4] 2> 1558897 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1558911 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34477.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558918 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34477.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558918 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34477.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1558919 INFO (jetty-launcher-3068-thread-3) [n:127.0.0.1:34477_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node3/.
[junit4] 2> 1559115 INFO (jetty-launcher-3068-thread-4) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41683/solr
[junit4] 2> 1559116 INFO (zkConnectionManagerCallback-3125-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1559117 INFO (zkConnectionManagerCallback-3127-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1559119 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 1559120 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1559121 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1559121 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:46571_solr
[junit4] 2> 1559121 INFO (zkCallback-3112-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1559121 INFO (zkCallback-3091-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1559121 INFO (zkCallback-3084-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1559121 INFO (zkCallback-3098-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1559122 INFO (zkCallback-3119-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1559122 INFO (zkCallback-3126-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1559130 INFO (zkCallback-3105-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1559139 INFO (zkConnectionManagerCallback-3134-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1559139 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1559140 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41683/solr ready
[junit4] 2> 1559140 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1559154 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46571.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1559162 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46571.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1559162 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46571.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1559163 INFO (jetty-launcher-3068-thread-4) [n:127.0.0.1:46571_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node4/.
[junit4] 2> 1559197 INFO (zkConnectionManagerCallback-3137-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1559199 INFO (zkConnectionManagerCallback-3142-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1559200 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1559200 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41683/solr ready
[junit4] 2> 1559250 INFO (qtp28109472-13218) [n:127.0.0.1:40599_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :overseerstatus with params action=OVERSEERSTATUS&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1559254 INFO (qtp28109472-13218) [n:127.0.0.1:40599_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={action=OVERSEERSTATUS&wt=javabin&version=2} status=0 QTime=4
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 1559317 WARN (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 1559330 WARN (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1559332 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1559348 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_localdomain_36407_hdfs____jtnjrm/webapp
[junit4] 2> 1559878 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:36407
[junit4] 2> 1559967 WARN (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1559968 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1559982 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_43449_datanode____af8czl/webapp
[junit4] 2> 1560465 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:43449
[junit4] 2> 1560511 WARN (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1560512 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1560522 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_39041_datanode____ij84by/webapp
[junit4] 2> 1560704 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:32923) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1560716 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x99296bfce55ce: from storage DS-eb3fda9c-e7ba-4acf-8024-7a4af041a4b1 node DatanodeRegistration(127.0.0.1:35021, datanodeUuid=d08af0f7-ce86-46c2-a068-d271f98f3566, infoPort=39327, infoSecurePort=0, ipcPort=40443, storageInfo=lv=-56;cid=testClusterID;nsid=667750826;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 1560716 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x99296bfce55ce: from storage DS-46abd5c8-6968-47ce-9706-b6813a4de15f node DatanodeRegistration(127.0.0.1:35021, datanodeUuid=d08af0f7-ce86-46c2-a068-d271f98f3566, infoPort=39327, infoSecurePort=0, ipcPort=40443, storageInfo=lv=-56;cid=testClusterID;nsid=667750826;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1560980 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:39041
[junit4] 2> 1561085 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:32923) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1561091 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x99296d6382e62: from storage DS-8588abe1-cb50-4ad3-9ba6-61538d22569c node DatanodeRegistration(127.0.0.1:46119, datanodeUuid=56d89570-dd5d-473f-9a10-98de9df0e00c, infoPort=37053, infoSecurePort=0, ipcPort=43007, storageInfo=lv=-56;cid=testClusterID;nsid=667750826;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 1561091 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x99296d6382e62: from storage DS-aa572dc5-bb4a-4201-afe2-0374efa76068 node DatanodeRegistration(127.0.0.1:46119, datanodeUuid=56d89570-dd5d-473f-9a10-98de9df0e00c, infoPort=37053, infoSecurePort=0, ipcPort=43007, storageInfo=lv=-56;cid=testClusterID;nsid=667750826;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1561256 INFO (OverseerCollectionConfigSetProcessor-72234177385791494-127.0.0.1:40599_solr-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 1561330 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1561332 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 1561332 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1561332 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1561332 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1561333 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@735bf9{/solr,null,AVAILABLE}
[junit4] 2> 1561333 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@1ff0492{SSL,[ssl, http/1.1]}{127.0.0.1:46543}
[junit4] 2> 1561333 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.e.j.s.Server Started @1561364ms
[junit4] 2> 1561333 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=46543}
[junit4] 2> 1561333 ERROR (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1561333 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1561334 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1561334 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1561334 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1561334 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T16:48:59.886Z
[junit4] 2> 1561335 INFO (zkConnectionManagerCallback-3146-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1561335 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1562469 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41683/solr
[junit4] 2> 1562470 INFO (zkConnectionManagerCallback-3150-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562472 INFO (zkConnectionManagerCallback-3152-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562477 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1562478 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1562480 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1562480 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:46543_solr
[junit4] 2> 1562484 INFO (zkCallback-3091-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562484 INFO (zkCallback-3098-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562484 INFO (zkCallback-3105-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562484 INFO (zkCallback-3112-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562484 INFO (zkCallback-3084-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562484 INFO (zkCallback-3141-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562484 INFO (zkCallback-3126-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562485 INFO (zkCallback-3119-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562485 INFO (zkCallback-3141-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562485 INFO (zkCallback-3133-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562492 INFO (zkCallback-3151-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1562501 INFO (zkConnectionManagerCallback-3159-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562501 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (5)
[junit4] 2> 1562502 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41683/solr ready
[junit4] 2> 1562502 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1562515 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46543.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1562522 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46543.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1562522 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46543.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1562523 INFO (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[6AEEF24E2527C591]) [n:127.0.0.1:46543_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node5/.
[junit4] 2> 1562580 INFO (zkConnectionManagerCallback-3162-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562629 INFO (qtp2684070-13197) [n:127.0.0.1:34477_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&autoAddReplicas=false&name=MoveReplicaHDFSTest_failed_coll_true&nrtReplicas=2&action=CREATE&numShards=2&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1562630 INFO (OverseerThreadFactory-6523-thread-2) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection MoveReplicaHDFSTest_failed_coll_true
[junit4] 2> 1562734 INFO (OverseerStateUpdate-72234177385791494-127.0.0.1:40599_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_failed_coll_true",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:40189/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1562735 INFO (OverseerStateUpdate-72234177385791494-127.0.0.1:40599_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_failed_coll_true",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:40599/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1562736 INFO (OverseerStateUpdate-72234177385791494-127.0.0.1:40599_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_failed_coll_true",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:46543/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1562736 INFO (OverseerStateUpdate-72234177385791494-127.0.0.1:40599_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_failed_coll_true",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:46571/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1562964 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1562964 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1563016 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1563016 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1563020 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1563998 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1564002 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] Schema name=minimal
[junit4] 2> 1564004 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1564004 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
[junit4] 2> 1564005 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40599.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n2' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1564005 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:32923/data
[junit4] 2> 1564005 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1564005 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1564005 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node1/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2], dataDir=[hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data/]
[junit4] 2> 1564006 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data/snapshot_metadata
[junit4] 2> 1564022 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data
[junit4] 2> 1564031 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1564031 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1564031 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1564051 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] Schema name=minimal
[junit4] 2> 1564051 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1564052 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] Schema name=minimal
[junit4] 2> 1564053 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1564053 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1564053 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
[junit4] 2> 1564053 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
[junit4] 2> 1564053 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1564053 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
[junit4] 2> 1564055 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40189.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n1' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1564055 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46543.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n4' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1564055 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46571.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n6' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1df29e
[junit4] 2> 1564055 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:32923/data
[junit4] 2> 1564055 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:32923/data
[junit4] 2> 1564055 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:32923/data
[junit4] 2> 1564055 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1564056 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1564056 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1564056 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1564056 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1564056 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1564056 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node5/MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4], dataDir=[hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data/]
[junit4] 2> 1564056 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node4/MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6], dataDir=[hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data/]
[junit4] 2> 1564056 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-001/node2/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1], dataDir=[hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data/]
[junit4] 2> 1564057 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data/snapshot_metadata
[junit4] 2> 1564057 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data/snapshot_metadata
[junit4] 2> 1564057 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data/snapshot_metadata
[junit4] 2> 1564062 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data/index
[junit4] 2> 1564071 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data
[junit4] 2> 1564075 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data
[junit4] 2> 1564081 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data
[junit4] 2> 1564099 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data/index
[junit4] 2> 1564109 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data/index
[junit4] 2> 1564114 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:32923/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data/index
[junit4] 2> 1564196 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46119 is added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-46abd5c8-6968-47ce-9706-b6813a4de15f:NORMAL:127.0.0.1:35021|RBW], ReplicaUC[[DISK]DS-8588abe1-cb50-4ad3-9ba6-61538d22569c:NORMAL:127.0.0.1:46119|FINALIZED]]} size 0
[junit4] 2> 1564197 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:35021 is added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-8588abe1-cb50-4ad3-9ba6-61538d22569c:NORMAL:127.0.0.1:46119|FINALIZED], ReplicaUC[[DISK]DS-eb3fda9c-e7ba-4acf-8024-7a4af041a4b1:NORMAL:127.0.0.1:35021|FINALIZED]]} size 0
[junit4] 2> 1564237 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46119 is added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-46abd5c8-6968-47ce-9706-b6813a4de15f:NORMAL:127.0.0.1:35021|RBW], ReplicaUC[[DISK]DS-aa572dc5-bb4a-4201-afe2-0374efa76068:NORMAL:127.0.0.1:46119|FINALIZED]]} size 0
[junit4] 2> 1564238 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:35021 is added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-46abd5c8-6968-47ce-9706-b6813a4de15f:NORMAL:127.0.0.1:35021|RBW], ReplicaUC[[DISK]DS-aa572dc5-bb4a-4201-afe2-0374efa76068:NORMAL:127.0.0.1:46119|FINALIZED]]} size 0
[junit4] 2> 1564266 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:35021 is added to blk_1073741827_1003{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-8588abe1-cb50-4ad3-9ba6-61538d22569c:NORMAL:127.0.0.1:46119|RBW], ReplicaUC[[DISK]DS-eb3fda9c-e7ba-4acf-8024-7a4af041a4b1:NORMAL:127.0.0.1:35021|RBW]]} size 0
[junit4] 2> 1564267 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46119 is added to blk_1073741827_1003 size 69
[junit4] 2> 1564303 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:35021 is added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-aa572dc5-bb4a-4201-afe2-0374efa76068:NORMAL:127.0.0.1:46119|RBW], ReplicaUC[[DISK]DS-46abd5c8-6968-47ce-9706-b6813a4de15f:NORMAL:127.0.0.1:35021|FINALIZED]]} size 0
[junit4] 2> 1564304 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46119 is added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-aa572dc5-bb4a-4201-afe2-0374efa76068:NORMAL:127.0.0.1:46119|RBW], ReplicaUC[[DISK]DS-46abd5c8-6968-47ce-9706-b6813a4de15f:NORMAL:127.0.0.1:35021|FINALIZED]]} size 0
[junit4] 2> 1564315 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1564315 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1564315 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1564355 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1564355 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1564368 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1564368 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1564368 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1564372 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1564372 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1564372 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1564383 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1564383 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1564385 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1564385 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1564390 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@1242d68[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] main]
[junit4] 2> 1564395 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1564395 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1564396 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1564397 INFO (searcherExecutor-6558-thread-1-processing-n:127.0.0.1:40599_solr x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2 c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] Registered new searcher Searcher@1242d68[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1564397 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603810983096090624
[junit4] 2> 1564399 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1564399 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1564399 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1564402 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard1 to Terms{values={core_node5=0}, version=0}
[junit4] 2> 1564405 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard1: total=2 found=1 timeoutin=9999ms
[junit4] 2> 1564406 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@6e4238[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] main]
[junit4] 2> 1564407 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1564407 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1564408 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1564408 INFO (searcherExecutor-6559-thread-1-processing-n:127.0.0.1:46571_solr x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6 c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] Registered new searcher Searcher@6e4238[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1564408 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603810983107624960
[junit4] 2> 1564411 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard2 to Terms{values={core_node8=0}, version=0}
[junit4] 2> 1564413 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard2: total=2 found=1 timeoutin=9999ms
[junit4] 2> 1564416 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1564416 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1564419 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@1d3a8c6[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] main]
[junit4] 2> 1564420 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1564421 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1564422 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1564424 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603810983124402176
[junit4] 2> 1564425 INFO (searcherExecutor-6561-thread-1-processing-n:127.0.0.1:46543_solr x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4 c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] Registered new searcher Searcher@1d3a8c6[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1564427 INFO (qtp6976172-13518) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard2 to Terms{values={core_node7=0, core_node8=0}, version=1}
[junit4] 2> 1564437 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@15dba82[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] main]
[junit4] 2> 1564438 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1564438 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1564439 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1564440 INFO (searcherExecutor-6560-thread-1-processing-n:127.0.0.1:40189_solr x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1 c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] Registered new searcher Searcher@15dba82[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1564440 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603810983141179392
[junit4] 2> 1564443 INFO (qtp8453744-13211) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard1 to Terms{values={core_node3=0, core_node5=0}, version=1}
[junit4] 2> 1564906 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1564906 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1564906 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:40599/solr/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2/
[junit4] 2> 1564907 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2 url=https://127.0.0.1:40599/solr START replicas=[https://127.0.0.1:40189/solr/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1/] nUpdates=100
[junit4] 2> 1564907 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2 url=https://127.0.0.1:40599/solr DONE. We have no versions. sync failed.
[junit4] 2> 1564914 INFO (qtp8453744-13190) [n:127.0.0.1:40189_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.S.Request [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=1
[junit4] 2> 1564914 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
[junit4] 2> 1564914 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
[junit4] 2> 1564914 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1564914 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1564914 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1564915 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:46571/solr/MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6/
[junit4] 2> 1564915 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6 url=https://127.0.0.1:46571/solr START replicas=[https://127.0.0.1:46543/solr/MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4/] nUpdates=100
[junit4] 2> 1564915 INFO (qtp13284131-13202) [n:127.0.0.1:46571_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6 url=https://127.0.0.1:46571/solr DONE. We have no versions. sync failed.
[junit4] 2> 1564916 INFO (qtp28109472-13220) [n:127.0.0.1:40599_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:40599/solr/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2/ shard1
[junit4] 2> 1564922 INFO (qtp6976172-13613) [n:127.0.0.1:46543_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.S.Request [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&v
[...truncated too long message...]
en called
[junit4] 2> 1616935 INFO (org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5bfbe6) [ ] BlockStateChange BLOCK* BlockManager: ask 127.0.0.1:35021 to delete [blk_1073742016_1193, blk_1073742017_1194, blk_1073742018_1195, blk_1073742019_1196, blk_1073742020_1197, blk_1073742006_1183, blk_1073742007_1184, blk_1073742008_1185, blk_1073742009_1186, blk_1073742010_1187, blk_1073742011_1188, blk_1073742012_1189, blk_1073742013_1190, blk_1073742014_1191, blk_1073742015_1192]
[junit4] 2> 1616989 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0
[junit4] 2> 1617092 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:32923) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 1617092 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:32923) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-35959254-88.99.242.108-1529513337825 (Datanode Uuid 56d89570-dd5d-473f-9a10-98de9df0e00c) service to localhost.localdomain/127.0.0.1:32923
[junit4] 2> 1617095 WARN (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called
[junit4] 2> 1617104 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0
[junit4] 2> 1617107 WARN (7151988@qtp-17149323-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:43449) [ ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is false. Rechecking.
[junit4] 2> 1617107 WARN (7151988@qtp-17149323-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:43449) [ ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is false
[junit4] 2> 1617207 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:32923) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 1617207 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:32923) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-35959254-88.99.242.108-1529513337825 (Datanode Uuid d08af0f7-ce86-46c2-a068-d271f98f3566) service to localhost.localdomain/127.0.0.1:32923
[junit4] 2> 1617215 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:0
[junit4] 2> 1617344 INFO (SUITE-MoveReplicaHDFSTest-seed#[6AEEF24E2527C591]-worker) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:41683 41683
[junit4] 2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_6AEEF24E2527C591-001
[junit4] 2> Jun 20, 2018 4:49:55 PM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 66 leaked thread(s).
[junit4] 2> NOTE: test params are: codec=Lucene70, sim=RandomSimilarity(queryNorm=false): {}, locale=ar, timezone=Etc/GMT-14
[junit4] 2> NOTE: Linux 4.13.0-41-generic i386/Oracle Corporation 1.8.0_172 (32-bit)/cpus=8,threads=2,free=258393800,total=467664896
[junit4] 2> NOTE: All tests run in this JVM: [BlockJoinFacetSimpleTest, TestLockTree, MaxSizeAutoCommitTest, TriggerCooldownIntegrationTest, BlobRepositoryCloudTest, NotRequiredUniqueKeyTest, TestJmxIntegration, ShowFileRequestHandlerTest, NumberUtilsTest, TestAtomicUpdateErrorCases, LIROnShardRestartTest, CheckHdfsIndexTest, CoreAdminHandlerTest, TestTlogReplica, TestUseDocValuesAsStored, TestShardHandlerFactory, SolrJmxReporterCloudTest, AnalyticsQueryTest, HdfsUnloadDistributedZkTest, ParsingFieldUpdateProcessorsTest, TestClusterStateMutator, CollectionsAPISolrJTest, SpellCheckCollatorWithCollapseTest, PingRequestHandlerTest, DocExpirationUpdateProcessorFactoryTest, OverseerStatusTest, TestClusterStateProvider, TestIntervalFaceting, SolrMetricsIntegrationTest, TestRestManager, HDFSCollectionsAPITest, TestInPlaceUpdatesDistrib, TestRequestForwarding, LeaderElectionIntegrationTest, TestLazyCores, TestNestedDocsSort, TestHdfsCloudBackupRestore, TestNumericTerms64, TestSimDistributedQueue, SmileWriterTest, SolrCloudReportersTest, SolrCoreMetricManagerTest, TestTolerantSearch, NodeMutatorTest, TestFieldSortValues, TestJsonRequest, FullHLLTest, OverseerCollectionConfigSetProcessorTest, TestPrepRecovery, TestExportWriter, TestFieldCacheWithThreads, ResourceLoaderTest, TestFaceting, TestHashQParserPlugin, BasicDistributedZk2Test, TestRangeQuery, TolerantUpdateProcessorTest, TestXmlQParser, TestSolr4Spatial2, HdfsChaosMonkeyNothingIsSafeTest, TestCloudJSONFacetSKG, TestDocTermOrdsUninvertLimit, TestSuggestSpellingConverter, TestReRankQParserPlugin, TestLuceneIndexBackCompat, TestLeaderElectionWithEmptyReplica, AssignTest, MultiTermTest, TestQueryTypes, DocValuesTest, TestRandomDVFaceting, TestSolrCoreProperties, TemplateUpdateProcessorTest, TestMinMaxOnMultiValuedField, TestSolrDeletionPolicy2, ClusterStateTest, TestRawTransformer, TestDistributedMap, TestLRUStatsCache, DistributedSpellCheckComponentTest, DistributedQueueTest, TestSearcherReuse, TestRealTimeGet, TestPerFieldSimilarityWithDefaultOverride, SuggestComponentTest, BigEndianAscendingWordSerializerTest, TestBackupRepositoryFactory, DistributedFacetExistsSmallTest, RegexBytesRefFilterTest, TestFastLRUCache, CdcrReplicationHandlerTest, RequestLoggingTest, FileBasedSpellCheckerTest, TestNodeAddedTrigger, CoreAdminCreateDiscoverTest, SolrCmdDistributorTest, TestManagedStopFilterFactory, TestStressCloudBlindAtomicUpdates, LegacyCloudClusterPropTest, XmlInterpolationTest, TestConfigSetsAPI, V2ApiIntegrationTest, ClusterStateUpdateTest, MissingSegmentRecoveryTest, TestExecutePlanAction, RegexBoostProcessorTest, CoreMergeIndexesAdminHandlerTest, EnumFieldTest, TestEmbeddedSolrServerSchemaAPI, SchemaApiFailureTest, MetricsHistoryHandlerTest, TestInitParams, SpatialFilterTest, CdcrBidirectionalTest, TestMacroExpander, TestDFRSimilarityFactory, IndexSchemaRuntimeFieldTest, DistribCursorPagingTest, RecoveryZkTest, TestChildDocTransformer, TestClassicSimilarityFactory, TestRestoreCore, TestCollationFieldDocValues, TestQuerySenderListener, UpdateLogTest, TestCoreContainer, TestSolrXml, ReturnFieldsTest, CollectionsAPIAsyncDistributedZkTest, BlockCacheTest, ImplicitSnitchTest, TestNumericRangeQuery32, TestRemoteStreaming, DistanceFunctionTest, SolrInfoBeanTest, CacheHeaderTest, TestSurroundQueryParser, LukeRequestHandlerTest, TestReversedWildcardFilterFactory, TermsComponentTest, TestIndexingPerformance, TestSolrQueryParser, LoggingHandlerTest, TestCollationField, MBeansHandlerTest, UpdateParamsTest, TestQuerySenderNoQuery, CopyFieldTest, ResponseLogComponentTest, SolrIndexConfigTest, BadComponentTest, TestStressRecovery, CloudExitableDirectoryReaderTest, ConcurrentCreateRoutedAliasTest, ConfigSetsAPITest, ConnectionManagerTest, DeleteInactiveReplicaTest, DeleteStatusTest, ForceLeaderTest, MoveReplicaHDFSTest]
[junit4] Completed [581/820 (1!)] on J2 in 68.92s, 4 tests, 1 error, 1 skipped <<< FAILURES!
[...truncated 47281 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2162/consoleText
[repro] Revision: 3b0edb0d667dbfa8c8ffb6c836a68a6f07effc00
[repro] Ant options: "-Dargs=-server -XX:+UseParallelGC"
[repro] Repro line: ant test -Dtestcase=MoveReplicaHDFSTest -Dtests.method=testFailedMove -Dtests.seed=6AEEF24E2527C591 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ar -Dtests.timezone=Etc/GMT-14 -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[repro] ant clean
[...truncated 6 lines...]
[repro] Test suites by module:
[repro] solr/core
[repro] MoveReplicaHDFSTest
[repro] ant compile-test
[...truncated 3318 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.MoveReplicaHDFSTest" -Dtests.showOutput=onerror "-Dargs=-server -XX:+UseParallelGC" -Dtests.seed=6AEEF24E2527C591 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ar -Dtests.timezone=Etc/GMT-14 -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[...truncated 86 lines...]
[repro] Failures:
[repro] 0/5 failed: org.apache.solr.cloud.MoveReplicaHDFSTest
[repro] Exiting with code 0
[...truncated 40 lines...]
[JENKINS] Lucene-Solr-7.x-Linux (64bit/jdk-10.0.1) - Build # 2165 -
Still Unstable!
Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2165/
Java: 64bit/jdk-10.0.1 -XX:+UseCompressedOops -XX:+UseParallelGC
1 tests failed.
FAILED: org.apache.solr.cloud.MoveReplicaHDFSTest.testFailedMove
Error Message:
No live SolrServers available to handle this request:[https://127.0.0.1:42151/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:43191/solr/MoveReplicaHDFSTest_failed_coll_true]
Stack Trace:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request:[https://127.0.0.1:42151/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:43191/solr/MoveReplicaHDFSTest_failed_coll_true]
at __randomizedtesting.SeedInfo.seed([53A22E86C22B5FB5:F96FFD7475F88A65]:0)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:462)
at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1106)
at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:886)
at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:993)
at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:819)
at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:942)
at org.apache.solr.cloud.MoveReplicaTest.testFailedMove(MoveReplicaTest.java:288)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:844)
Caused by: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at https://127.0.0.1:42151/solr/MoveReplicaHDFSTest_failed_coll_true: no servers hosting shard: shard1
at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:436)
... 45 more
Build Log:
[...truncated 14361 lines...]
[junit4] Suite: org.apache.solr.cloud.MoveReplicaHDFSTest
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/init-core-data-001
[junit4] 2> 1491341 WARN (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=2 numCloses=2
[junit4] 2> 1491342 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 1491342 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", ssl=0.0/0.0, value=0.0/0.0, clientAuth=0.0/0.0)
[junit4] 2> 1491343 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> 1491343 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 4 servers in /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001
[junit4] 2> 1491343 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1491344 INFO (Thread-2848) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1491344 INFO (Thread-2848) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 1491345 ERROR (Thread-2848) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 1491444 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.c.ZkTestServer start zk server on port:41695
[junit4] 2> 1491445 INFO (zkConnectionManagerCallback-7430-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-2) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-3) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-1) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-4) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-4) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-4) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-4) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1491447 INFO (jetty-launcher-7427-thread-1) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@5da75ad9{/solr,null,AVAILABLE}
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@608f7c60{/solr,null,AVAILABLE}
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.e.j.s.AbstractConnector Started ServerConnector@6677c0d{SSL,[ssl, http/1.1]}{127.0.0.1:42151}
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@4e40c9ea{SSL,[ssl, http/1.1]}{127.0.0.1:33537}
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.e.j.s.Server Started @1491475ms
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.e.j.s.Server Started @1491475ms
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=33537}
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=42151}
[junit4] 2> 1491448 ERROR (jetty-launcher-7427-thread-4) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1491448 ERROR (jetty-launcher-7427-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T05:40:26.510766Z
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1491448 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T05:40:26.510794Z
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-2) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-2) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-2) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-3) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-3) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-3) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-2) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3493f9ec{/solr,null,AVAILABLE}
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-2) [ ] o.e.j.s.AbstractConnector Started ServerConnector@7ebbb9fe{SSL,[ssl, http/1.1]}{127.0.0.1:33185}
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-3) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@489ca8a5{/solr,null,AVAILABLE}
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-2) [ ] o.e.j.s.Server Started @1491476ms
[junit4] 2> 1491449 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=33185}
[junit4] 2> 1491449 ERROR (jetty-launcher-7427-thread-2) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T05:40:26.512086Z
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.e.j.s.AbstractConnector Started ServerConnector@6087f5db{SSL,[ssl, http/1.1]}{127.0.0.1:43191}
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.e.j.s.Server Started @1491477ms
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=43191}
[junit4] 2> 1491450 ERROR (jetty-launcher-7427-thread-3) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1491450 INFO (zkConnectionManagerCallback-7433-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T05:40:26.512399Z
[junit4] 2> 1491450 INFO (zkConnectionManagerCallback-7434-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1491450 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1491451 INFO (zkConnectionManagerCallback-7436-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491451 INFO (zkConnectionManagerCallback-7438-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491451 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1491451 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1491534 INFO (jetty-launcher-7427-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41695/solr
[junit4] 2> 1491541 INFO (zkConnectionManagerCallback-7442-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491554 INFO (jetty-launcher-7427-thread-2) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41695/solr
[junit4] 2> 1491597 INFO (zkConnectionManagerCallback-7448-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491601 INFO (zkConnectionManagerCallback-7445-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491626 INFO (zkConnectionManagerCallback-7450-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491631 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1491631 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:33185_solr
[junit4] 2> 1491631 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.c.Overseer Overseer (id=72237211049459720-127.0.0.1:33185_solr-n_0000000000) starting
[junit4] 2> 1491646 INFO (zkConnectionManagerCallback-7459-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491647 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41695/solr ready
[junit4] 2> 1491657 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1491658 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:33537_solr
[junit4] 2> 1491665 INFO (OverseerStateUpdate-72237211049459720-127.0.0.1:33185_solr-n_0000000000) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1491666 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1491666 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:33185_solr
[junit4] 2> 1491678 INFO (zkCallback-7449-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1491686 INFO (zkCallback-7458-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1491686 INFO (zkCallback-7444-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1491686 INFO (zkCallback-7444-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1491698 INFO (zkConnectionManagerCallback-7465-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491698 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1491699 INFO (jetty-launcher-7427-thread-3) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41695/solr
[junit4] 2> 1491699 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41695/solr ready
[junit4] 2> 1491701 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1491709 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1491709 INFO (zkConnectionManagerCallback-7469-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491715 INFO (zkConnectionManagerCallback-7472-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491716 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33185.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491718 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1491719 INFO (jetty-launcher-7427-thread-4) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41695/solr
[junit4] 2> 1491719 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1491720 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33537.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491723 INFO (zkConnectionManagerCallback-7478-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491724 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1491724 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:43191_solr
[junit4] 2> 1491724 INFO (zkConnectionManagerCallback-7480-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491725 INFO (zkCallback-7458-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1491725 INFO (zkCallback-7449-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1491725 INFO (zkCallback-7444-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1491725 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33185.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491725 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33185.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491728 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33537.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491728 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 1491729 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33537.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491730 INFO (jetty-launcher-7427-thread-1) [n:127.0.0.1:33537_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node1/.
[junit4] 2> 1491730 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1491734 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1491734 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:42151_solr
[junit4] 2> 1491734 INFO (zkCallback-7449-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1491735 INFO (zkCallback-7444-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1491735 INFO (zkCallback-7458-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1491736 INFO (zkCallback-7464-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1491737 INFO (jetty-launcher-7427-thread-2) [n:127.0.0.1:33185_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node2/.
[junit4] 2> 1491744 INFO (zkCallback-7464-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1491746 INFO (zkCallback-7471-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (4)
[junit4] 2> 1491746 INFO (zkCallback-7479-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1491748 INFO (zkConnectionManagerCallback-7492-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491749 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1491749 INFO (zkConnectionManagerCallback-7487-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491749 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41695/solr ready
[junit4] 2> 1491750 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1491750 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41695/solr ready
[junit4] 2> 1491753 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1491753 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1491769 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_42151.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491771 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43191.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491773 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_42151.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491774 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_42151.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491774 INFO (jetty-launcher-7427-thread-4) [n:127.0.0.1:42151_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node4/.
[junit4] 2> 1491778 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43191.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491778 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43191.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1491779 INFO (jetty-launcher-7427-thread-3) [n:127.0.0.1:43191_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node3/.
[junit4] 2> 1491821 INFO (zkConnectionManagerCallback-7496-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491823 INFO (zkConnectionManagerCallback-7501-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1491823 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1491824 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41695/solr ready
[junit4] 2> 1491836 INFO (qtp637897816-20548) [n:127.0.0.1:33185_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :overseerstatus with params action=OVERSEERSTATUS&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1491840 INFO (qtp637897816-20548) [n:127.0.0.1:33185_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={action=OVERSEERSTATUS&wt=javabin&version=2} status=0 QTime=3
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 1492400 WARN (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 1492512 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Logging to org.apache.logging.slf4j.Log4jLogger@6e36aa53 via org.mortbay.log.Slf4jLog
[junit4] 2> 1492525 WARN (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1492563 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1492594 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_localdomain_34789_hdfs____.hbm7jj/webapp
[junit4] 2> 1492694 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:34789
[junit4] 2> 1493098 WARN (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1493100 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1493103 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_41259_datanode____.uwym00/webapp
[junit4] 2> 1493175 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41259
[junit4] 2> 1493191 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] i.n.u.i.PlatformDependent Your platform does not provide complete low-level API for accessing direct buffers reliably. Unless explicitly requested, heap buffer will always be preferred to avoid potential system unstability.
[junit4] 2> 1493318 WARN (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1493319 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1493324 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_36081_datanode____af46a5/webapp
[junit4] 2> 1493401 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:36081
[junit4] 2> 1493837 INFO (OverseerCollectionConfigSetProcessor-72237211049459720-127.0.0.1:33185_solr-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 1494086 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:34081) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1494113 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:34081) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1494141 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9bcb076734535: from storage DS-9f6f59df-1809-40e9-979a-b62ddf111ac7 node DatanodeRegistration(127.0.0.1:33505, datanodeUuid=eefc25b0-9fc6-469b-b01d-e46f78864f18, infoPort=41599, infoSecurePort=0, ipcPort=46295, storageInfo=lv=-56;cid=testClusterID;nsid=1287633184;c=0), blocks: 0, hasStaleStorage: true, processing time: 1 msecs
[junit4] 2> 1494141 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9bcb076941a72: from storage DS-ffcfaf84-762a-4bc9-9085-2ceeccc1410c node DatanodeRegistration(127.0.0.1:33787, datanodeUuid=82dd1a1c-471b-43f7-be03-5a00c40f6a5e, infoPort=37373, infoSecurePort=0, ipcPort=42077, storageInfo=lv=-56;cid=testClusterID;nsid=1287633184;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 1494141 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9bcb076734535: from storage DS-d900f31d-e217-4231-8e18-a04467facfa3 node DatanodeRegistration(127.0.0.1:33505, datanodeUuid=eefc25b0-9fc6-469b-b01d-e46f78864f18, infoPort=41599, infoSecurePort=0, ipcPort=46295, storageInfo=lv=-56;cid=testClusterID;nsid=1287633184;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1494145 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9bcb076941a72: from storage DS-6735df42-23a2-4ea6-aabb-d83775a4a42b node DatanodeRegistration(127.0.0.1:33787, datanodeUuid=82dd1a1c-471b-43f7-be03-5a00c40f6a5e, infoPort=37373, infoSecurePort=0, ipcPort=42077, storageInfo=lv=-56;cid=testClusterID;nsid=1287633184;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1494350 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.SolrTestCaseJ4 ###Starting testNormalMove
[junit4] 2> 1494351 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1494353 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 1494361 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1494361 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1494362 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1494362 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@7a844a7b{/solr,null,AVAILABLE}
[junit4] 2> 1494362 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@76fcd97d{SSL,[ssl, http/1.1]}{127.0.0.1:40247}
[junit4] 2> 1494362 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.e.j.s.Server Started @1494389ms
[junit4] 2> 1494362 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=40247}
[junit4] 2> 1494363 ERROR (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1494363 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1494363 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 1494363 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1494363 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1494363 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T05:40:29.425432Z
[junit4] 2> 1494365 INFO (zkConnectionManagerCallback-7505-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1494366 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1494509 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41695/solr
[junit4] 2> 1494510 INFO (zkConnectionManagerCallback-7509-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1494512 INFO (zkConnectionManagerCallback-7511-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1494515 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 1494516 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1494517 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1494517 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:40247_solr
[junit4] 2> 1494518 INFO (zkCallback-7444-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494518 INFO (zkCallback-7479-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494518 INFO (zkCallback-7464-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494518 INFO (zkCallback-7449-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494518 INFO (zkCallback-7458-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494521 INFO (zkCallback-7471-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494535 INFO (zkCallback-7500-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494535 INFO (zkCallback-7510-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494535 INFO (zkCallback-7491-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494535 INFO (zkCallback-7500-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494536 INFO (zkCallback-7486-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 1494537 INFO (zkConnectionManagerCallback-7518-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1494538 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (5)
[junit4] 2> 1494538 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41695/solr ready
[junit4] 2> 1494538 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1494553 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40247.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1494565 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40247.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1494566 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40247.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1494567 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [n:127.0.0.1:40247_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node5/.
[junit4] 2> 1494627 INFO (zkConnectionManagerCallback-7521-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1494628 INFO (TEST-MoveReplicaHDFSTest.testNormalMove-seed#[53A22E86C22B5FB5]) [ ] o.a.s.c.MoveReplicaTest total_jettys: 5
[junit4] 2> 1494629 INFO (qtp637897816-20552) [n:127.0.0.1:33185_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&maxShardsPerNode=2&autoAddReplicas=false&name=MoveReplicaHDFSTest_coll_false&nrtReplicas=2&action=CREATE&numShards=2&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1494631 INFO (OverseerThreadFactory-6822-thread-2) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection MoveReplicaHDFSTest_coll_false
[junit4] 2> 1494736 INFO (OverseerStateUpdate-72237211049459720-127.0.0.1:33185_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_false",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_false_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:40247/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1494737 INFO (OverseerStateUpdate-72237211049459720-127.0.0.1:33185_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_false",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_false_shard1_replica_n2",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:43191/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1494738 INFO (OverseerStateUpdate-72237211049459720-127.0.0.1:33185_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_false",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_false_shard2_replica_n4",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:33537/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1494740 INFO (OverseerStateUpdate-72237211049459720-127.0.0.1:33185_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_false",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_false_shard2_replica_n7",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:42151/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1494994 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node6&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_false_shard2_replica_n4&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_false&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1494994 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_false_shard1_replica_n2&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_false&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1494994 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1494995 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_false_shard2_replica_n7&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_false&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1494996 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_false_shard1_replica_n1&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_false&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1496006 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1496006 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1496010 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_false_shard2_replica_n4] Schema name=minimal
[junit4] 2> 1496011 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1496011 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_false_shard2_replica_n4' using configuration from collection MoveReplicaHDFSTest_coll_false, trusted=true
[junit4] 2> 1496011 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_false_shard1_replica_n2] Schema name=minimal
[junit4] 2> 1496011 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33537.solr.core.MoveReplicaHDFSTest_coll_false.shard2.replica_n4' (registry 'solr.core.MoveReplicaHDFSTest_coll_false.shard2.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1496012 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1496012 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_false_shard1_replica_n2' using configuration from collection MoveReplicaHDFSTest_coll_false, trusted=true
[junit4] 2> 1496013 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43191.solr.core.MoveReplicaHDFSTest_coll_false.shard1.replica_n2' (registry 'solr.core.MoveReplicaHDFSTest_coll_false.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1496017 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1496029 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_false_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1496029 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:34081/data
[junit4] 2> 1496029 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:34081/data
[junit4] 2> 1496029 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1496029 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1496029 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1496029 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1496029 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_false_shard1_replica_n2] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node3/MoveReplicaHDFSTest_coll_false_shard1_replica_n2], dataDir=[hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node5/data/]
[junit4] 2> 1496029 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_false_shard2_replica_n4] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node1/MoveReplicaHDFSTest_coll_false_shard2_replica_n4], dataDir=[hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node6/data/]
[junit4] 2> 1496030 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 1496030 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1496030 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_false_shard1_replica_n1' using configuration from collection MoveReplicaHDFSTest_coll_false, trusted=true
[junit4] 2> 1496030 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_40247.solr.core.MoveReplicaHDFSTest_coll_false.shard1.replica_n1' (registry 'solr.core.MoveReplicaHDFSTest_coll_false.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1496031 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:34081/data
[junit4] 2> 1496031 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1496031 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1496031 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_false_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node5/MoveReplicaHDFSTest_coll_false_shard1_replica_n1], dataDir=[hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node3/data/]
[junit4] 2> 1496031 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node6/data/snapshot_metadata
[junit4] 2> 1496031 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node5/data/snapshot_metadata
[junit4] 2> 1496032 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node3/data/snapshot_metadata
[junit4] 2> 1496033 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_false_shard2_replica_n7] Schema name=minimal
[junit4] 2> 1496034 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1496034 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_false_shard2_replica_n7' using configuration from collection MoveReplicaHDFSTest_coll_false, trusted=true
[junit4] 2> 1496034 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_42151.solr.core.MoveReplicaHDFSTest_coll_false.shard2.replica_n7' (registry 'solr.core.MoveReplicaHDFSTest_coll_false.shard2.replica_n7') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a4a9152
[junit4] 2> 1496034 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:34081/data
[junit4] 2> 1496034 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1496034 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1496034 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_false_shard2_replica_n7] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-001/node4/MoveReplicaHDFSTest_coll_false_shard2_replica_n7], dataDir=[hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node8/data/]
[junit4] 2> 1496036 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node8/data/snapshot_metadata
[junit4] 2> 1496066 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node6/data
[junit4] 2> 1496066 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node3/data
[junit4] 2> 1496066 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node8/data
[junit4] 2> 1496066 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node5/data
[junit4] 2> 1496085 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node6/data/index
[junit4] 2> 1496088 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node8/data/index
[junit4] 2> 1496089 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node3/data/index
[junit4] 2> 1496091 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:34081/data/MoveReplicaHDFSTest_coll_false/core_node5/data/index
[junit4] 2> 1496213 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33505 is added to blk_1073741827_1003{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6735df42-23a2-4ea6-aabb-d83775a4a42b:NORMAL:127.0.0.1:33787|RBW], ReplicaUC[[DISK]DS-9f6f59df-1809-40e9-979a-b62ddf111ac7:NORMAL:127.0.0.1:33505|RBW]]} size 69
[junit4] 2> 1496213 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33787 is added to blk_1073741825_1001{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-9f6f59df-1809-40e9-979a-b62ddf111ac7:NORMAL:127.0.0.1:33505|RBW], ReplicaUC[[DISK]DS-6735df42-23a2-4ea6-aabb-d83775a4a42b:NORMAL:127.0.0.1:33787|RBW]]} size 69
[junit4] 2> 1496213 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33505 is added to blk_1073741825_1001 size 69
[junit4] 2> 1496213 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33505 is added to blk_1073741826_1002{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-d900f31d-e217-4231-8e18-a04467facfa3:NORMAL:127.0.0.1:33505|RBW], ReplicaUC[[DISK]DS-6735df42-23a2-4ea6-aabb-d83775a4a42b:NORMAL:127.0.0.1:33787|RBW]]} size 69
[junit4] 2> 1496213 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33505 is added to blk_1073741828_1004{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-d900f31d-e217-4231-8e18-a04467facfa3:NORMAL:127.0.0.1:33505|RBW], ReplicaUC[[DISK]DS-6735df42-23a2-4ea6-aabb-d83775a4a42b:NORMAL:127.0.0.1:33787|RBW]]} size 69
[junit4] 2> 1496214 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33787 is added to blk_1073741827_1003 size 69
[junit4] 2> 1496214 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33787 is added to blk_1073741826_1002 size 69
[junit4] 2> 1496214 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33787 is added to blk_1073741828_1004 size 69
[junit4] 2> 1496682 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1496682 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1496682 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1496692 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1496692 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1496692 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1496695 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1496695 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1496702 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1496702 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1496710 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1496710 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1496710 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1496710 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1496710 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1496710 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1496722 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1496722 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1496725 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1496725 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1496750 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@267d30c9[MoveReplicaHDFSTest_coll_false_shard2_replica_n4] main]
[junit4] 2> 1496751 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1496751 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1496753 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1496753 INFO (searcherExecutor-6854-thread-1-processing-n:127.0.0.1:33537_solr x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4 c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_false_shard2_replica_n4] Registered new searcher Searcher@267d30c9[MoveReplicaHDFSTest_coll_false_shard2_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1496754 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603859520491094016
[junit4] 2> 1496764 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@3c470dc4[MoveReplicaHDFSTest_coll_false_shard1_replica_n2] main]
[junit4] 2> 1496770 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1496770 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1496771 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1496771 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603859520508919808
[junit4] 2> 1496771 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_false/terms/shard2 to Terms{values={core_node6=0}, version=0}
[junit4] 2> 1496772 INFO (searcherExecutor-6855-thread-1-processing-n:127.0.0.1:43191_solr x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2 c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_false_shard1_replica_n2] Registered new searcher Searcher@3c470dc4[MoveReplicaHDFSTest_coll_false_shard1_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1496774 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard2: total=2 found=1 timeoutin=9999ms
[junit4] 2> 1496775 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_false/terms/shard1 to Terms{values={core_node5=0}, version=0}
[junit4] 2> 1496778 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.s.SolrIndexSearcher Opening [Searcher@8db51c7[MoveReplicaHDFSTest_coll_false_shard2_replica_n7] main]
[junit4] 2> 1496778 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@53fe0a7[MoveReplicaHDFSTest_coll_false_shard1_replica_n1] main]
[junit4] 2> 1496781 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1496782 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1496782 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard1: total=2 found=1 timeoutin=9999ms
[junit4] 2> 1496782 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1496782 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1496783 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1496783 INFO (searcherExecutor-6856-thread-1-processing-n:127.0.0.1:40247_solr x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1 c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_false_shard1_replica_n1] Registered new searcher Searcher@53fe0a7[MoveReplicaHDFSTest_coll_false_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1496783 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603859520521502720
[junit4] 2> 1496783 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1496784 INFO (searcherExecutor-6857-thread-1-processing-n:127.0.0.1:42151_solr x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7 c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_false_shard2_replica_n7] Registered new searcher Searcher@8db51c7[MoveReplicaHDFSTest_coll_false_shard2_replica_n7] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1496784 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603859520522551296
[junit4] 2> 1496785 INFO (qtp1061019398-20864) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_false/terms/shard1 to Terms{values={core_node3=0, core_node5=0}, version=1}
[junit4] 2> 1496786 INFO (qtp1902961938-20541) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_false/terms/shard2 to Terms{values={core_node6=0, core_node8=0}, version=1}
[junit4] 2> 1497275 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1497275 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1497275 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:33537/solr/MoveReplicaHDFSTest_coll_false_shard2_replica_n4/
[junit4] 2> 1497275 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_false_shard2_replica_n4 url=https://127.0.0.1:33537/solr START replicas=[https://127.0.0.1:42151/solr/MoveReplicaHDFSTest_coll_false_shard2_replica_n7/] nUpdates=100
[junit4] 2> 1497276 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_false_shard2_replica_n4 url=https://127.0.0.1:33537/solr DONE. We have no versions. sync failed.
[junit4] 2> 1497279 INFO (qtp1902961938-20545) [n:127.0.0.1:42151_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n7] o.a.s.c.S.Request [MoveReplicaHDFSTest_coll_false_shard2_replica_n7] webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 1497279 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
[junit4] 2> 1497279 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
[junit4] 2> 1497279 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1497282 INFO (qtp1840536124-20544) [n:127.0.0.1:33537_solr c:MoveReplicaHDFSTest_coll_false s:shard2 r:core_node6 x:MoveReplicaHDFSTest_coll_false_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:33537/solr/MoveReplicaHDFSTest_coll_false_shard2_replica_n4/ shard2
[junit4] 2> 1497283 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1497283 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1497283 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43191/solr/MoveReplicaHDFSTest_coll_false_shard1_replica_n2/
[junit4] 2> 1497283 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_false_shard1_replica_n2 url=https://127.0.0.1:43191/solr START replicas=[https://127.0.0.1:40247/solr/MoveReplicaHDFSTest_coll_false_shard1_replica_n1/] nUpdates=100
[junit4] 2> 1497285 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_false_shard1_replica_n2 url=https://127.0.0.1:43191/solr DONE. We have no versions. sync failed.
[junit4] 2> 1497289 INFO (qtp1061019398-20866) [n:127.0.0.1:40247_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node3 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n1] o.a.s.c.S.Request [MoveReplicaHDFSTest_coll_false_shard1_replica_n1] webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 1497289 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
[junit4] 2> 1497289 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
[junit4] 2> 1497289 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1497291 INFO (qtp1143895146-20553) [n:127.0.0.1:43191_solr c:MoveReplicaHDFSTest_coll_false s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_false_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43191/s
[...truncated too long message...]
data watches above limit:
[junit4] 2>
[junit4] 2> 17 /solr/collections/MoveReplicaHDFSTest_coll_false/state.json
[junit4] 2> 14 /solr/clusterprops.json
[junit4] 2> 14 /solr/clusterstate.json
[junit4] 2> 7 /solr/collections/MoveReplicaHDFSTest_coll_true/state.json
[junit4] 2> 4 /solr/collections/MoveReplicaHDFSTest_failed_coll_true/state.json
[junit4] 2> 2 /solr/collections/MoveReplicaHDFSTest_coll_false/leader_elect/shard1/election/72237211049459733-core_node3-n_0000000001
[junit4] 2> 2 /solr/collections/MoveReplicaHDFSTest_coll_true/leader_elect/shard2/election/72237211049459724-core_node8-n_0000000001
[junit4] 2> 2 /solr/autoscaling.json
[junit4] 2>
[junit4] 2> Maximum concurrent children watches above limit:
[junit4] 2>
[junit4] 2> 25 /solr/collections
[junit4] 2> 18 /solr/live_nodes
[junit4] 2> 2 /solr/overseer/queue
[junit4] 2> 2 /solr/autoscaling/events/.scheduled_maintenance
[junit4] 2> 2 /solr/autoscaling/events/.auto_add_replicas
[junit4] 2> 2 /solr/overseer/collection-queue-work
[junit4] 2> 2 /solr/overseer/queue-work
[junit4] 2>
[junit4] 2> 1548559 WARN (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called
[junit4] 2> 1548680 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0
[junit4] 2> 1548781 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:34081) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 1548781 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:34081) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-1549254781-88.99.242.108-1529559627181 (Datanode Uuid eefc25b0-9fc6-469b-b01d-e46f78864f18) service to localhost.localdomain/127.0.0.1:34081
[junit4] 2> 1548783 WARN (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called
[junit4] 2> 1548801 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0
[junit4] 2> 1548902 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:34081) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 1548902 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:34081) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-1549254781-88.99.242.108-1529559627181 (Datanode Uuid 82dd1a1c-471b-43f7-be03-5a00c40f6a5e) service to localhost.localdomain/127.0.0.1:34081
[junit4] 2> 1548911 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:0
[junit4] 2> 1549022 INFO (SUITE-MoveReplicaHDFSTest-seed#[53A22E86C22B5FB5]-worker) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:41695 41695
[junit4] 2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_53A22E86C22B5FB5-001
[junit4] 2> Jun 21, 2018 5:41:24 AM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 67 leaked thread(s).
[junit4] 2> NOTE: test params are: codec=Lucene70, sim=RandomSimilarity(queryNorm=true): {}, locale=bez, timezone=Europe/Skopje
[junit4] 2> NOTE: Linux 4.13.0-41-generic amd64/Oracle Corporation 10.0.1 (64-bit)/cpus=8,threads=6,free=301727760,total=522715136
[junit4] 2> NOTE: All tests run in this JVM: [TestGraphMLResponseWriter, AnalyticsMergeStrategyTest, TestFoldingMultitermQuery, LeaderVoteWaitTimeoutTest, TestPayloadCheckQParserPlugin, MetricsHandlerTest, SolrCloudExampleTest, RulesTest, HdfsTlogReplayBufferedWhileIndexingTest, DistribJoinFromCollectionTest, TestSegmentSorting, DOMUtilTest, TestCustomSort, TimeRoutedAliasUpdateProcessorTest, TestSolrConfigHandler, TestReload, TriggerSetPropertiesIntegrationTest, TestDocTermOrds, ReplaceNodeNoTargetTest, TestScoreJoinQPScore, CloneFieldUpdateProcessorFactoryTest, TestStressLiveNodes, IndexSchemaTest, TestReloadAndDeleteDocs, CollectionTooManyReplicasTest, ConcurrentDeleteAndCreateCollectionTest, ShardSplitTest, AutoscalingHistoryHandlerTest, SolrRrdBackendFactoryTest, TestBM25SimilarityFactory, MetricsConfigTest, DirectUpdateHandlerOptimizeTest, SyncSliceTest, ZkShardTermsTest, SolrCoreTest, PeerSyncReplicationTest, TestUnInvertedFieldException, DeleteNodeTest, IgnoreLargeDocumentProcessorFactoryTest, TestDynamicLoading, PreAnalyzedFieldManagedSchemaCloudTest, BufferStoreTest, TestCloudPseudoReturnFields, ChangedSchemaMergeTest, BigEndianAscendingWordDeserializerTest, SimpleMLTQParserTest, IndexSizeTriggerTest, PrimitiveFieldTypeTest, TermsComponentTest, TestSQLHandlerNonCloud, RandomizedTaggerTest, TestCorePropertiesReload, CreateRoutedAliasTest, TestPolicyCloud, DirectSolrConnectionTest, TestPullReplicaErrorHandling, TestExclusionRuleCollectionAccess, NodeAddedTriggerIntegrationTest, SolrInfoBeanTest, UtilsToolTest, BadComponentTest, ConnectionReuseTest, CachingDirectoryFactoryTest, TestSubQueryTransformerCrossCore, TestNumericTokenStream, TestFuzzyAnalyzedSuggestions, PrimUtilsTest, DeleteStatusTest, TestZkAclsWithHadoopAuth, TestReversedWildcardFilterFactory, CdcrReplicationDistributedZkTest, TestCloudSchemaless, ExplicitHLLTest, TestSolrCloudWithSecureImpersonation, TestGraphTermsQParserPlugin, StressHdfsTest, TestUnifiedSolrHighlighter, DocValuesMissingTest, TestFastOutputStream, TestSafeXMLParsing, TestSolrIndexConfig, PeerSyncTest, ChaosMonkeyNothingIsSafeWithPullReplicasTest, SearchHandlerTest, SortByFunctionTest, DistributedIntervalFacetingTest, TestRawResponseWriter, TestSurroundQueryParser, TestEmbeddedSolrServerConstructors, TestDistribDocBasedVersion, TestSystemCollAutoCreate, TestRebalanceLeaders, TransactionLogTest, ConfigSetsAPITest, TestReplicationHandler, TriLevelCompositeIdRoutingTest, TestDistributedGrouping, TestCloudConsistency, AddBlockUpdateTest, TestDistribStateManager, TestCharFilters, SpatialRPTFieldTypeTest, AlternateDirectoryTest, SearchRateTriggerIntegrationTest, TestHighFrequencyDictionaryFactory, ReplicaListTransformerTest, AliasIntegrationTest, TestLegacyFieldReuse, TestCollectionsAPIViaSolrCloudCluster, TestSolrQueryResponse, TestRecovery, TestPolicyCloud, TestSolrCoreSnapshots, DistanceUnitsTest, SolrSlf4jReporterTest, TestSortByMinMaxFunction, AddSchemaFieldsUpdateProcessorFactoryTest, TestStressRecovery, TestCopyFieldCollectionResource, TestLeaderInitiatedRecoveryThread, TestInfoStreamLogging, TestElisionMultitermQuery, TestTrackingShardHandlerFactory, NodeLostTriggerTest, PolyFieldTest, TestCoreBackup, ScheduledMaintenanceTriggerTest, TestBlendedInfixSuggestions, HighlighterTest, OverseerRolesTest, TestNonDefinedSimilarityFactory, UUIDFieldTest, MoveReplicaHDFSTest]
[junit4] Completed [566/820 (1!)] on J1 in 67.26s, 4 tests, 1 error, 1 skipped <<< FAILURES!
[...truncated 44875 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2165/consoleText
[repro] Revision: 3b0edb0d667dbfa8c8ffb6c836a68a6f07effc00
[repro] Ant options: "-Dargs=-XX:+UseCompressedOops -XX:+UseParallelGC"
[repro] Repro line: ant test -Dtestcase=MoveReplicaHDFSTest -Dtests.method=testFailedMove -Dtests.seed=53A22E86C22B5FB5 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bez -Dtests.timezone=Europe/Skopje -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[repro] ant clean
[...truncated 6 lines...]
[repro] Test suites by module:
[repro] solr/core
[repro] MoveReplicaHDFSTest
[repro] ant compile-test
[...truncated 3323 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.MoveReplicaHDFSTest" -Dtests.showOutput=onerror "-Dargs=-XX:+UseCompressedOops -XX:+UseParallelGC" -Dtests.seed=53A22E86C22B5FB5 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bez -Dtests.timezone=Europe/Skopje -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[...truncated 91 lines...]
[repro] Failures:
[repro] 0/5 failed: org.apache.solr.cloud.MoveReplicaHDFSTest
[repro] Exiting with code 0
[...truncated 40 lines...]
[JENKINS] Lucene-Solr-7.x-Linux (64bit/jdk-10.0.1) - Build # 2164 -
Still Unstable!
Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2164/
Java: 64bit/jdk-10.0.1 -XX:-UseCompressedOops -XX:+UseSerialGC
1 tests failed.
FAILED: org.apache.solr.cloud.MoveReplicaHDFSTest.testFailedMove
Error Message:
No live SolrServers available to handle this request:[https://127.0.0.1:34583/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:33395/solr/MoveReplicaHDFSTest_failed_coll_true]
Stack Trace:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request:[https://127.0.0.1:34583/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:33395/solr/MoveReplicaHDFSTest_failed_coll_true]
at __randomizedtesting.SeedInfo.seed([2FF014368F98092F:853DC7C4384BDCFF]:0)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:462)
at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1106)
at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:886)
at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:993)
at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:819)
at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:942)
at org.apache.solr.cloud.MoveReplicaTest.testFailedMove(MoveReplicaTest.java:288)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:844)
Caused by: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at https://127.0.0.1:34583/solr/MoveReplicaHDFSTest_failed_coll_true: no servers hosting shard: shard2
at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:436)
... 45 more
Build Log:
[...truncated 12812 lines...]
[junit4] Suite: org.apache.solr.cloud.MoveReplicaHDFSTest
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/init-core-data-001
[junit4] 2> 197252 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 4 servers in /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001
[junit4] 2> 197253 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 197253 INFO (Thread-778) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 197253 INFO (Thread-778) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 197264 ERROR (Thread-778) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 197353 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.s.c.ZkTestServer start zk server on port:45817
[junit4] 2> 197354 INFO (zkConnectionManagerCallback-773-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197357 INFO (jetty-launcher-770-thread-2) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 197357 INFO (jetty-launcher-770-thread-4) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 197357 INFO (jetty-launcher-770-thread-3) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 197357 INFO (jetty-launcher-770-thread-1) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 197358 INFO (jetty-launcher-770-thread-3) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 197358 INFO (jetty-launcher-770-thread-3) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 197358 INFO (jetty-launcher-770-thread-3) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 197358 INFO (jetty-launcher-770-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 197358 INFO (jetty-launcher-770-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 197358 INFO (jetty-launcher-770-thread-1) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 197358 INFO (jetty-launcher-770-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@5c95f98f{/solr,null,AVAILABLE}
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-3) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@315465ee{/solr,null,AVAILABLE}
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-2) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-2) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-2) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-4) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-4) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-4) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 197359 INFO (jetty-launcher-770-thread-2) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3322413b{/solr,null,AVAILABLE}
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.e.j.s.AbstractConnector Started ServerConnector@7e4b26f9{SSL,[ssl, http/1.1]}{127.0.0.1:33395}
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.e.j.s.AbstractConnector Started ServerConnector@11df3fef{SSL,[ssl, http/1.1]}{127.0.0.1:34583}
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.e.j.s.Server Started @197387ms
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.e.j.s.Server Started @197387ms
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=33395}
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=34583}
[junit4] 2> 197360 ERROR (jetty-launcher-770-thread-3) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 197360 ERROR (jetty-launcher-770-thread-2) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T00:55:07.139575Z
[junit4] 2> 197360 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T00:55:07.139582Z
[junit4] 2> 197365 INFO (jetty-launcher-770-thread-4) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@11aa579f{/solr,null,AVAILABLE}
[junit4] 2> 197365 INFO (jetty-launcher-770-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@47942e2f{SSL,[ssl, http/1.1]}{127.0.0.1:44759}
[junit4] 2> 197365 INFO (jetty-launcher-770-thread-1) [ ] o.e.j.s.Server Started @197393ms
[junit4] 2> 197365 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=44759}
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.e.j.s.AbstractConnector Started ServerConnector@23d44e62{SSL,[ssl, http/1.1]}{127.0.0.1:41797}
[junit4] 2> 197366 ERROR (jetty-launcher-770-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.e.j.s.Server Started @197393ms
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 197366 INFO (zkConnectionManagerCallback-776-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=41797}
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 197366 INFO (zkConnectionManagerCallback-777-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T00:55:07.145192Z
[junit4] 2> 197366 ERROR (jetty-launcher-770-thread-4) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 197366 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T00:55:07.145441Z
[junit4] 2> 197380 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100a2571fa00001, likely client has closed socket
[junit4] 2> 197380 INFO (zkConnectionManagerCallback-781-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197380 INFO (zkConnectionManagerCallback-779-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197381 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 197381 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 197434 INFO (jetty-launcher-770-thread-2) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45817/solr
[junit4] 2> 197435 INFO (zkConnectionManagerCallback-785-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197440 INFO (zkConnectionManagerCallback-787-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197458 INFO (jetty-launcher-770-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45817/solr
[junit4] 2> 197459 INFO (zkConnectionManagerCallback-793-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197460 INFO (zkConnectionManagerCallback-795-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197465 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 197466 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:44759_solr
[junit4] 2> 197466 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.c.Overseer Overseer (id=72236089114361864-127.0.0.1:44759_solr-n_0000000000) starting
[junit4] 2> 197472 INFO (zkConnectionManagerCallback-802-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197473 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45817/solr ready
[junit4] 2> 197477 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:44759_solr
[junit4] 2> 197487 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 197488 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 197489 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 197489 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34583_solr
[junit4] 2> 197493 INFO (zkCallback-794-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 197501 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 197509 INFO (zkCallback-801-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 197509 INFO (zkCallback-786-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 197512 INFO (zkConnectionManagerCallback-808-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197513 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 197513 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_44759.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197513 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45817/solr ready
[junit4] 2> 197516 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 197519 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_44759.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197519 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_44759.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197520 INFO (jetty-launcher-770-thread-1) [n:127.0.0.1:44759_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node1/.
[junit4] 2> 197521 INFO (jetty-launcher-770-thread-3) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45817/solr
[junit4] 2> 197523 INFO (zkConnectionManagerCallback-813-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197524 INFO (zkConnectionManagerCallback-815-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197526 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 197527 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 197527 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34583.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197528 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 197528 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:33395_solr
[junit4] 2> 197532 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34583.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197532 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34583.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197532 INFO (zkCallback-786-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 197532 INFO (zkCallback-794-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 197532 INFO (zkCallback-801-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 197533 INFO (zkCallback-814-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 197533 INFO (jetty-launcher-770-thread-2) [n:127.0.0.1:34583_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node2/.
[junit4] 2> 197533 INFO (zkCallback-807-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 197541 INFO (zkConnectionManagerCallback-822-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197541 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 197542 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45817/solr ready
[junit4] 2> 197542 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 197551 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33395.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197555 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33395.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197555 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33395.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197556 INFO (jetty-launcher-770-thread-3) [n:127.0.0.1:33395_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node3/.
[junit4] 2> 197600 INFO (jetty-launcher-770-thread-4) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45817/solr
[junit4] 2> 197601 INFO (zkConnectionManagerCallback-827-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197603 INFO (zkConnectionManagerCallback-829-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197606 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 197607 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 197608 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 197608 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:41797_solr
[junit4] 2> 197608 INFO (zkCallback-786-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 197608 INFO (zkCallback-794-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 197608 INFO (zkCallback-814-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 197608 INFO (zkCallback-801-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 197608 INFO (zkCallback-807-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 197609 INFO (zkCallback-821-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 197609 INFO (zkCallback-828-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 197618 INFO (zkConnectionManagerCallback-836-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197618 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 197619 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45817/solr ready
[junit4] 2> 197619 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 197626 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_41797.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197630 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_41797.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197630 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_41797.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 197631 INFO (jetty-launcher-770-thread-4) [n:127.0.0.1:41797_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node4/.
[junit4] 2> 197661 INFO (zkConnectionManagerCallback-839-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197663 INFO (zkConnectionManagerCallback-844-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 197664 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 197664 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45817/solr ready
[junit4] 2> 197676 INFO (qtp2125805845-3321) [n:127.0.0.1:44759_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :overseerstatus with params action=OVERSEERSTATUS&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 197680 INFO (qtp2125805845-3321) [n:127.0.0.1:44759_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={action=OVERSEERSTATUS&wt=javabin&version=2} status=0 QTime=4
[junit4] 2> 197917 WARN (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.h.u.NativeCodeLoader Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 198339 WARN (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 198450 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Logging to org.apache.logging.slf4j.Log4jLogger@3834be5a via org.mortbay.log.Slf4jLog
[junit4] 2> 198463 WARN (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 198602 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 198625 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_localdomain_34261_hdfs____lr056u/webapp
[junit4] 2> 198718 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:34261
[junit4] 2> 199087 WARN (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 199089 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 199091 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_41601_datanode____7di329/webapp
[junit4] 2> 199163 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41601
[junit4] 2> 199176 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] i.n.u.i.PlatformDependent Your platform does not provide complete low-level API for accessing direct buffers reliably. Unless explicitly requested, heap buffer will always be preferred to avoid potential system unstability.
[junit4] 2> 199299 WARN (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 199301 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 199308 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_40589_datanode____r8ohcx/webapp
[junit4] 2> 199393 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:40589
[junit4] 2> 199678 INFO (OverseerCollectionConfigSetProcessor-72236089114361864-127.0.0.1:44759_solr-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 199988 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:38085) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 199988 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:38085) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 200027 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9ad1e8b118142: from storage DS-b541d187-a1f4-4194-89ab-d8bbc46c5852 node DatanodeRegistration(127.0.0.1:45239, datanodeUuid=a22e3ce2-7243-41cf-a24b-68133fbf6c1a, infoPort=33433, infoSecurePort=0, ipcPort=36939, storageInfo=lv=-56;cid=testClusterID;nsid=1424552063;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 200028 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9ad1e8b1180e0: from storage DS-e4f0e60e-0ebe-4e1d-97a5-c89b9c28fdc0 node DatanodeRegistration(127.0.0.1:43381, datanodeUuid=c565ecfa-a7e6-4bad-a303-3ca58448773c, infoPort=34213, infoSecurePort=0, ipcPort=36727, storageInfo=lv=-56;cid=testClusterID;nsid=1424552063;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 200028 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9ad1e8b118142: from storage DS-f8809ccb-40c3-49d2-a851-9228344bf768 node DatanodeRegistration(127.0.0.1:45239, datanodeUuid=a22e3ce2-7243-41cf-a24b-68133fbf6c1a, infoPort=33433, infoSecurePort=0, ipcPort=36939, storageInfo=lv=-56;cid=testClusterID;nsid=1424552063;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 200028 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x9ad1e8b1180e0: from storage DS-1bdf89e5-03dc-4b82-ad89-01409859a7b1 node DatanodeRegistration(127.0.0.1:43381, datanodeUuid=c565ecfa-a7e6-4bad-a303-3ca58448773c, infoPort=34213, infoSecurePort=0, ipcPort=36727, storageInfo=lv=-56;cid=testClusterID;nsid=1424552063;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 200144 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 200145 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 200146 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 200146 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 200147 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 200147 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@61470bc3{/solr,null,AVAILABLE}
[junit4] 2> 200147 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@7c30b8a6{SSL,[ssl, http/1.1]}{127.0.0.1:35185}
[junit4] 2> 200148 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.e.j.s.Server Started @200175ms
[junit4] 2> 200148 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=35185}
[junit4] 2> 200148 ERROR (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 200148 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 200148 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 200148 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 200148 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 200148 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-21T00:55:09.927628Z
[junit4] 2> 200153 INFO (zkConnectionManagerCallback-848-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 200153 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 200157 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100a2571fa00013, likely client has closed socket
[junit4] 2> 200233 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45817/solr
[junit4] 2> 200234 INFO (zkConnectionManagerCallback-852-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 200235 INFO (zkConnectionManagerCallback-854-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 200238 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 200239 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 200240 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 200240 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:35185_solr
[junit4] 2> 200241 INFO (zkCallback-794-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200241 INFO (zkCallback-801-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200241 INFO (zkCallback-828-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200241 INFO (zkCallback-814-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200241 INFO (zkCallback-786-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200241 INFO (zkCallback-821-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200241 INFO (zkCallback-807-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200243 INFO (zkCallback-843-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200243 INFO (zkCallback-835-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200247 INFO (zkCallback-843-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200247 INFO (zkCallback-853-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
[junit4] 2> 200277 INFO (zkConnectionManagerCallback-861-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 200277 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (5)
[junit4] 2> 200278 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45817/solr ready
[junit4] 2> 200280 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 200302 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_35185.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 200316 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_35185.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 200316 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_35185.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 200318 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [n:127.0.0.1:35185_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node5/.
[junit4] 2> 200375 INFO (zkConnectionManagerCallback-864-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 200376 INFO (TEST-MoveReplicaHDFSTest.test-seed#[2FF014368F98092F]) [ ] o.a.s.c.MoveReplicaTest total_jettys: 5
[junit4] 2> 200391 INFO (qtp732471074-3310) [n:127.0.0.1:33395_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&maxShardsPerNode=2&autoAddReplicas=false&name=MoveReplicaHDFSTest_coll_true&nrtReplicas=2&action=CREATE&numShards=2&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 200396 INFO (OverseerThreadFactory-1665-thread-2) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection MoveReplicaHDFSTest_coll_true
[junit4] 2> 200503 INFO (OverseerStateUpdate-72236089114361864-127.0.0.1:44759_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_true",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_true_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:35185/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 200508 INFO (OverseerStateUpdate-72236089114361864-127.0.0.1:44759_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_true",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_true_shard1_replica_n3",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:33395/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 200510 INFO (OverseerStateUpdate-72236089114361864-127.0.0.1:44759_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_true",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_true_shard2_replica_n4",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:41797/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 200512 INFO (OverseerStateUpdate-72236089114361864-127.0.0.1:44759_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"MoveReplicaHDFSTest_coll_true",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"MoveReplicaHDFSTest_coll_true_shard2_replica_n6",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:34583/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 200805 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_true_shard1_replica_n3&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_true&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 200816 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node2&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_true_shard1_replica_n1&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_true&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 200841 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_true_shard2_replica_n4&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_true&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 200861 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_true_shard2_replica_n6&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_true&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 201831 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 201836 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_true_shard1_replica_n3] Schema name=minimal
[junit4] 2> 201838 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 201838 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_true_shard1_replica_n3' using configuration from collection MoveReplicaHDFSTest_coll_true, trusted=true
[junit4] 2> 201839 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_33395.solr.core.MoveReplicaHDFSTest_coll_true.shard1.replica_n3' (registry 'solr.core.MoveReplicaHDFSTest_coll_true.shard1.replica_n3') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 201849 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:38085/data
[junit4] 2> 201849 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 201849 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 201849 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_true_shard1_replica_n3] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node3/MoveReplicaHDFSTest_coll_true_shard1_replica_n3], dataDir=[hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node5/data/]
[junit4] 2> 201849 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 201852 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node5/data/snapshot_metadata
[junit4] 2> 201868 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_true_shard1_replica_n1] Schema name=minimal
[junit4] 2> 201871 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 201871 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_true_shard1_replica_n1' using configuration from collection MoveReplicaHDFSTest_coll_true, trusted=true
[junit4] 2> 201871 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 201871 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_35185.solr.core.MoveReplicaHDFSTest_coll_true.shard1.replica_n1' (registry 'solr.core.MoveReplicaHDFSTest_coll_true.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 201872 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:38085/data
[junit4] 2> 201872 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 201872 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 201872 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_true_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node5/MoveReplicaHDFSTest_coll_true_shard1_replica_n1], dataDir=[hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node2/data/]
[junit4] 2> 201874 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node2/data/snapshot_metadata
[junit4] 2> 201883 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_true_shard2_replica_n4] Schema name=minimal
[junit4] 2> 201887 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 201893 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_coll_true_shard2_replica_n6] Schema name=minimal
[junit4] 2> 201895 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 201895 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_true_shard2_replica_n6' using configuration from collection MoveReplicaHDFSTest_coll_true, trusted=true
[junit4] 2> 201895 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34583.solr.core.MoveReplicaHDFSTest_coll_true.shard2.replica_n6' (registry 'solr.core.MoveReplicaHDFSTest_coll_true.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 201895 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:38085/data
[junit4] 2> 201895 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 201896 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 201896 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_true_shard2_replica_n6] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node2/MoveReplicaHDFSTest_coll_true_shard2_replica_n6], dataDir=[hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node8/data/]
[junit4] 2> 201897 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 201897 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_coll_true_shard2_replica_n4' using configuration from collection MoveReplicaHDFSTest_coll_true, trusted=true
[junit4] 2> 201897 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node8/data/snapshot_metadata
[junit4] 2> 201897 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_41797.solr.core.MoveReplicaHDFSTest_coll_true.shard2.replica_n4' (registry 'solr.core.MoveReplicaHDFSTest_coll_true.shard2.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@7c99ed52
[junit4] 2> 201897 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:38085/data
[junit4] 2> 201897 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 201897 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 201897 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_coll_true_shard2_replica_n4] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-001/node4/MoveReplicaHDFSTest_coll_true_shard2_replica_n4], dataDir=[hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node7/data/]
[junit4] 2> 201899 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node7/data/snapshot_metadata
[junit4] 2> 201934 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node2/data
[junit4] 2> 201935 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node8/data
[junit4] 2> 201941 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node5/data
[junit4] 2> 201969 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node7/data
[junit4] 2> 202046 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node5/data/index
[junit4] 2> 202086 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node7/data/index
[junit4] 2> 202093 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node2/data/index
[junit4] 2> 202104 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:38085/data/MoveReplicaHDFSTest_coll_true/core_node8/data/index
[junit4] 2> 202365 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:43381 is added to blk_1073741825_1001{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-1bdf89e5-03dc-4b82-ad89-01409859a7b1:NORMAL:127.0.0.1:43381|RBW], ReplicaUC[[DISK]DS-f8809ccb-40c3-49d2-a851-9228344bf768:NORMAL:127.0.0.1:45239|RBW]]} size 69
[junit4] 2> 202366 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:45239 is added to blk_1073741825_1001 size 69
[junit4] 2> 202366 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:43381 is added to blk_1073741826_1002{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-f8809ccb-40c3-49d2-a851-9228344bf768:NORMAL:127.0.0.1:45239|RBW], ReplicaUC[[DISK]DS-e4f0e60e-0ebe-4e1d-97a5-c89b9c28fdc0:NORMAL:127.0.0.1:43381|RBW]]} size 69
[junit4] 2> 202366 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:43381 is added to blk_1073741827_1003{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e4f0e60e-0ebe-4e1d-97a5-c89b9c28fdc0:NORMAL:127.0.0.1:43381|RBW], ReplicaUC[[DISK]DS-b541d187-a1f4-4194-89ab-d8bbc46c5852:NORMAL:127.0.0.1:45239|RBW]]} size 69
[junit4] 2> 202366 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:43381 is added to blk_1073741828_1004{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-1bdf89e5-03dc-4b82-ad89-01409859a7b1:NORMAL:127.0.0.1:43381|RBW], ReplicaUC[[DISK]DS-f8809ccb-40c3-49d2-a851-9228344bf768:NORMAL:127.0.0.1:45239|RBW]]} size 69
[junit4] 2> 202366 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:45239 is added to blk_1073741827_1003 size 69
[junit4] 2> 202366 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:45239 is added to blk_1073741826_1002 size 69
[junit4] 2> 202366 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:45239 is added to blk_1073741828_1004 size 69
[junit4] 2> 202891 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 202891 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 202891 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 202900 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 202900 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 202900 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 202905 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 202905 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 202914 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 202914 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 202933 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 202933 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 202934 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 202948 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 202948 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 202952 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 202952 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 202953 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 202965 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 202965 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 203006 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@3f738f8e[MoveReplicaHDFSTest_coll_true_shard2_replica_n4] main]
[junit4] 2> 203007 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 203008 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 203020 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 203021 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@22a31ec9[MoveReplicaHDFSTest_coll_true_shard1_replica_n3] main]
[junit4] 2> 203021 INFO (searcherExecutor-1700-thread-1-processing-n:127.0.0.1:41797_solr x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4 c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_true_shard2_replica_n4] Registered new searcher Searcher@3f738f8e[MoveReplicaHDFSTest_coll_true_shard2_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 203022 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603841569902821376
[junit4] 2> 203022 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 203022 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 203023 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 203023 INFO (searcherExecutor-1697-thread-1-processing-n:127.0.0.1:33395_solr x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3 c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_true_shard1_replica_n3] Registered new searcher Searcher@22a31ec9[MoveReplicaHDFSTest_coll_true_shard1_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 203024 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603841569904918528
[junit4] 2> 203026 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_true/terms/shard2 to Terms{values={core_node7=0}, version=0}
[junit4] 2> 203028 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_true/terms/shard1 to Terms{values={core_node5=0}, version=0}
[junit4] 2> 203029 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard2: total=2 found=1 timeoutin=9999ms
[junit4] 2> 203030 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard1: total=2 found=1 timeoutin=9999ms
[junit4] 2> 203035 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@2c0aafcf[MoveReplicaHDFSTest_coll_true_shard2_replica_n6] main]
[junit4] 2> 203035 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@675fe53e[MoveReplicaHDFSTest_coll_true_shard1_replica_n1] main]
[junit4] 2> 203036 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 203036 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 203036 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 203036 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 203037 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 203037 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 203038 INFO (searcherExecutor-1699-thread-1-processing-n:127.0.0.1:34583_solr x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6 c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_true_shard2_replica_n6] Registered new searcher Searcher@2c0aafcf[MoveReplicaHDFSTest_coll_true_shard2_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 203038 INFO (searcherExecutor-1698-thread-1-processing-n:127.0.0.1:35185_solr x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1 c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.SolrCore [MoveReplicaHDFSTest_coll_true_shard1_replica_n1] Registered new searcher Searcher@675fe53e[MoveReplicaHDFSTest_coll_true_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 203038 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603841569919598592
[junit4] 2> 203038 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603841569919598592
[junit4] 2> 203042 INFO (qtp51342802-3643) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_true/terms/shard1 to Terms{values={core_node2=0, core_node5=0}, version=1}
[junit4] 2> 203042 INFO (qtp285394463-3332) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_coll_true/terms/shard2 to Terms{values={core_node7=0, core_node8=0}, version=1}
[junit4] 2> 203530 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 203530 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 203530 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:41797/solr/MoveReplicaHDFSTest_coll_true_shard2_replica_n4/
[junit4] 2> 203530 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_true_shard2_replica_n4 url=https://127.0.0.1:41797/solr START replicas=[https://127.0.0.1:34583/solr/MoveReplicaHDFSTest_coll_true_shard2_replica_n6/] nUpdates=100
[junit4] 2> 203531 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_true_shard2_replica_n4 url=https://127.0.0.1:41797/solr DONE. We have no versions. sync failed.
[junit4] 2> 203531 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 203531 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 203531 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:33395/solr/MoveReplicaHDFSTest_coll_true_shard1_replica_n3/
[junit4] 2> 203532 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_true_shard1_replica_n3 url=https://127.0.0.1:33395/solr START replicas=[https://127.0.0.1:35185/solr/MoveReplicaHDFSTest_coll_true_shard1_replica_n1/] nUpdates=100
[junit4] 2> 203532 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_coll_true_shard1_replica_n3 url=https://127.0.0.1:33395/solr DONE. We have no versions. sync failed.
[junit4] 2> 203537 INFO (qtp285394463-3335) [n:127.0.0.1:34583_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n6] o.a.s.c.S.Request [MoveReplicaHDFSTest_coll_true_shard2_replica_n6] webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=2
[junit4] 2> 203538 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
[junit4] 2> 203538 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
[junit4] 2> 203538 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 203540 INFO (qtp51342802-3747) [n:127.0.0.1:35185_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node2 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n1] o.a.s.c.S.Request [MoveReplicaHDFSTest_coll_true_shard1_replica_n1] webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 203540 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
[junit4] 2> 203540 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
[junit4] 2> 203540 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 203546 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:41797/solr/MoveReplicaHDFSTest_coll_true_shard2_replica_n4/ shard2
[junit4] 2> 203547 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:33395/solr/MoveReplicaHDFSTest_coll_true_shard1_replica_n3/ shard1
[junit4] 2> 203648 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 203648 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 203721 INFO (qtp732471074-3328) [n:127.0.0.1:33395_solr c:MoveReplicaHDFSTest_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_coll_true_shard1_replica_n3] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_true_shard1_replica_n3&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_true&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2915
[junit4] 2> 203721 INFO (qtp57621383-3336) [n:127.0.0.1:41797_solr c:MoveReplicaHDFSTest_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_coll_true_shard2_replica_n4] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_coll_true_shard2_replica_n4&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_coll_true&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2880
[junit4] 2> 203750 INFO (zkCallback-828-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/MoveReplicaHDFSTest_coll_true/state.json] for collection [MoveReplicaHDFSTest_coll_true] has occurred - updating... (live nodes size: [5])
[junit4]
[...truncated too long message...]
.session node0 Stopped scavenging
[junit4] 2> 258964 ERROR (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 258965 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:45817 45817
[junit4] 2> 259093 INFO (Thread-778) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:45817 45817
[junit4] 2> 259093 WARN (Thread-778) [ ] o.a.s.c.ZkTestServer Watch limit violations:
[junit4] 2> Maximum concurrent create/delete watches above limit:
[junit4] 2>
[junit4] 2> 35 /solr/collections/MoveReplicaHDFSTest_coll_false/terms/shard1
[junit4] 2> 28 /solr/collections/MoveReplicaHDFSTest_coll_true/terms/shard1
[junit4] 2> 15 /solr/configs/conf1
[junit4] 2> 14 /solr/aliases.json
[junit4] 2> 10 /solr/collections/MoveReplicaHDFSTest_coll_false/terms/shard2
[junit4] 2> 10 /solr/collections/MoveReplicaHDFSTest_coll_true/terms/shard2
[junit4] 2> 9 /solr/collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard2
[junit4] 2> 7 /solr/collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard1
[junit4] 2> 5 /solr/security.json
[junit4] 2>
[junit4] 2> Maximum concurrent data watches above limit:
[junit4] 2>
[junit4] 2> 14 /solr/clusterprops.json
[junit4] 2> 14 /solr/clusterstate.json
[junit4] 2> 7 /solr/collections/MoveReplicaHDFSTest_coll_true/state.json
[junit4] 2> 7 /solr/collections/MoveReplicaHDFSTest_coll_false/state.json
[junit4] 2> 4 /solr/collections/MoveReplicaHDFSTest_failed_coll_true/state.json
[junit4] 2> 2 /solr/collections/MoveReplicaHDFSTest_coll_false/leader_elect/shard1/election/72236089114361868-core_node5-n_0000000001
[junit4] 2> 2 /solr/collections/MoveReplicaHDFSTest_coll_true/leader_elect/shard1/election/72236089114361877-core_node2-n_0000000001
[junit4] 2> 2 /solr/autoscaling.json
[junit4] 2> 2 /solr/overseer_elect/election/72236089114361868-127.0.0.1:33395_solr-n_0000000002
[junit4] 2>
[junit4] 2> Maximum concurrent children watches above limit:
[junit4] 2>
[junit4] 2> 25 /solr/collections
[junit4] 2> 18 /solr/live_nodes
[junit4] 2> 2 /solr/overseer/queue
[junit4] 2> 2 /solr/autoscaling/events/.scheduled_maintenance
[junit4] 2> 2 /solr/autoscaling/events/.auto_add_replicas
[junit4] 2> 2 /solr/overseer/collection-queue-work
[junit4] 2> 2 /solr/overseer/queue-work
[junit4] 2>
[junit4] 2> 259095 WARN (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called
[junit4] 2> 259233 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0
[junit4] 2> 259234 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:38085) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 259234 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:38085) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-538169291-88.99.242.108-1529542507833 (Datanode Uuid a22e3ce2-7243-41cf-a24b-68133fbf6c1a) service to localhost.localdomain/127.0.0.1:38085
[junit4] 2> 259236 WARN (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called
[junit4] 2> 259261 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0
[junit4] 2> 259265 WARN (1573925622@qtp-363750785-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41601) [ ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is false. Rechecking.
[junit4] 2> 259265 WARN (1573925622@qtp-363750785-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41601) [ ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is false
[junit4] 2> 259271 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:38085) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 259271 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:38085) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-538169291-88.99.242.108-1529542507833 (Datanode Uuid c565ecfa-a7e6-4bad-a303-3ca58448773c) service to localhost.localdomain/127.0.0.1:38085
[junit4] 2> 259309 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:0
[junit4] 2> 259405 INFO (SUITE-MoveReplicaHDFSTest-seed#[2FF014368F98092F]-worker) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:45817 45817
[junit4] 2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.MoveReplicaHDFSTest_2FF014368F98092F-001
[junit4] 2> Jun 21, 2018 12:56:09 AM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 67 leaked thread(s).
[junit4] 2> NOTE: test params are: codec=Asserting(Lucene70): {id=Lucene50(blocksize=128)}, docValues:{}, maxPointsInLeafNode=1799, maxMBSortInHeap=6.631717917404652, sim=RandomSimilarity(queryNorm=false): {}, locale=jmc, timezone=Europe/Riga
[junit4] 2> NOTE: Linux 4.13.0-41-generic amd64/Oracle Corporation 10.0.1 (64-bit)/cpus=8,threads=6,free=268702688,total=518979584
[junit4] 2> NOTE: All tests run in this JVM: [TestCursorMarkWithoutUniqueKey, HdfsRestartWhileUpdatingTest, LoggingHandlerTest, BadComponentTest, DataDrivenBlockJoinTest, TestPostingsSolrHighlighter, TestTrieFacet, DeleteNodeTest, TestConfigSetImmutable, ChaosMonkeyNothingIsSafeWithPullReplicasTest, CollectionTooManyReplicasTest, DistributedMLTComponentTest, TestFuzzyAnalyzedSuggestions, TestMultiValuedNumericRangeQuery, AddBlockUpdateTest, TestStressRecovery, CSVRequestHandlerTest, TestGraphMLResponseWriter, ClassificationUpdateProcessorFactoryTest, AddSchemaFieldsUpdateProcessorFactoryTest, TestDefaultStatsCache, TestLeaderInitiatedRecoveryThread, SchemaVersionSpecificBehaviorTest, AutoAddReplicasPlanActionTest, TestNumericTerms32, TestBM25SimilarityFactory, BlockJoinFacetRandomTest, TestCorePropertiesReload, ZkStateWriterTest, TestSimpleTrackingShardHandler, TestCopyFieldCollectionResource, DateMathParserTest, BasicAuthStandaloneTest, ConnectionReuseTest, PathHierarchyTokenizerFactoryTest, TestConfigSetsAPIZkFailure, ConjunctionSolrSpellCheckerTest, ExplicitHLLTest, CachingDirectoryFactoryTest, TestUniqueKeyFieldResource, TestZkChroot, OverseerTest, DistribJoinFromCollectionTest, DistributedFacetPivotLongTailTest, TestFieldCollectionResource, DocValuesNotIndexedTest, PrimUtilsTest, TestExpandComponent, MoveReplicaHDFSTest]
[junit4] Completed [128/820 (1!)] on J1 in 71.99s, 4 tests, 1 error, 1 skipped <<< FAILURES!
[...truncated 46410 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2164/consoleText
[repro] Revision: 3b0edb0d667dbfa8c8ffb6c836a68a6f07effc00
[repro] Ant options: "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC"
[repro] Repro line: ant test -Dtestcase=MoveReplicaHDFSTest -Dtests.method=testFailedMove -Dtests.seed=2FF014368F98092F -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=jmc -Dtests.timezone=Europe/Riga -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[repro] ant clean
[...truncated 6 lines...]
[repro] Test suites by module:
[repro] solr/core
[repro] MoveReplicaHDFSTest
[repro] ant compile-test
[...truncated 3323 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.MoveReplicaHDFSTest" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC" -Dtests.seed=2FF014368F98092F -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=jmc -Dtests.timezone=Europe/Riga -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[...truncated 91 lines...]
[repro] Failures:
[repro] 0/5 failed: org.apache.solr.cloud.MoveReplicaHDFSTest
[repro] Exiting with code 0
[...truncated 40 lines...]
[JENKINS] Lucene-Solr-7.x-Linux (64bit/jdk1.8.0_172) - Build # 2163
- Still Unstable!
Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2163/
Java: 64bit/jdk1.8.0_172 -XX:-UseCompressedOops -XX:+UseParallelGC
1 tests failed.
FAILED: org.apache.solr.update.processor.TestNamedUpdateProcessors.test
Error Message:
Error from server at https://127.0.0.1:40885/collection1: Async exception during distributed update: Error from server at https://127.0.0.1:44675/collection1_shard1_replica_n43: Bad Request request: https://127.0.0.1:44675/collection1_shard1_replica_n43/update?update.distrib=TOLEADER&distrib.from=https%3A%2F%2F127.0.0.1%3A40885%2Fcollection1_shard2_replica_n45%2F&wt=javabin&version=2 Remote error message: ERROR: [doc=123] multiple values encountered for non multiValued field test_s: [one, two]
Stack Trace:
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at https://127.0.0.1:40885/collection1: Async exception during distributed update: Error from server at https://127.0.0.1:44675/collection1_shard1_replica_n43: Bad Request
request: https://127.0.0.1:44675/collection1_shard1_replica_n43/update?update.distrib=TOLEADER&distrib.from=https%3A%2F%2F127.0.0.1%3A40885%2Fcollection1_shard2_replica_n45%2F&wt=javabin&version=2
Remote error message: ERROR: [doc=123] multiple values encountered for non multiValued field test_s: [one, two]
at __randomizedtesting.SeedInfo.seed([AB5C7C8C5A9F386C:23084356F4635594]:0)
at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
at org.apache.solr.client.solrj.SolrClient.add(SolrClient.java:173)
at org.apache.solr.client.solrj.SolrClient.add(SolrClient.java:138)
at org.apache.solr.client.solrj.SolrClient.add(SolrClient.java:152)
at org.apache.solr.update.processor.TestNamedUpdateProcessors.test(TestNamedUpdateProcessors.java:103)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:993)
at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:968)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
Build Log:
[...truncated 12457 lines...]
[junit4] Suite: org.apache.solr.update.processor.TestNamedUpdateProcessors
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/init-core-data-001
[junit4] 2> 49401 WARN (SUITE-TestNamedUpdateProcessors-seed#[AB5C7C8C5A9F386C]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=44 numCloses=44
[junit4] 2> 49401 INFO (SUITE-TestNamedUpdateProcessors-seed#[AB5C7C8C5A9F386C]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 49403 INFO (SUITE-TestNamedUpdateProcessors-seed#[AB5C7C8C5A9F386C]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason=, value=NaN, ssl=NaN, clientAuth=NaN)
[junit4] 2> 49403 INFO (SUITE-TestNamedUpdateProcessors-seed#[AB5C7C8C5A9F386C]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> 49403 INFO (SUITE-TestNamedUpdateProcessors-seed#[AB5C7C8C5A9F386C]-worker) [ ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /
[junit4] 2> 49418 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 49419 INFO (Thread-416) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 49419 INFO (Thread-416) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 49502 ERROR (Thread-416) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 49519 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.ZkTestServer start zk server on port:40523
[junit4] 2> 49597 INFO (zkConnectionManagerCallback-90-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 49630 INFO (zkConnectionManagerCallback-92-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 49643 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
[junit4] 2> 49658 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/schema-sql.xml to /configs/conf1/schema.xml
[junit4] 2> 49659 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 49662 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
[junit4] 2> 49664 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
[junit4] 2> 49666 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
[junit4] 2> 49668 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml
[junit4] 2> 49669 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
[junit4] 2> 49671 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 49672 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
[junit4] 2> 49674 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
[junit4] 2> 49698 INFO (zkConnectionManagerCallback-96-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 49716 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly asked otherwise
[junit4] 2> 49789 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 49818 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 49818 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 49819 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 49819 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@108f96aa{/,null,AVAILABLE}
[junit4] 2> 49823 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@70a5d30b{SSL,[ssl, http/1.1]}{127.0.0.1:39135}
[junit4] 2> 49823 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server Started @49854ms
[junit4] 2> 49823 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/tempDir-001/control/data, replicaType=NRT, hostContext=/, hostPort=39135, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/control-001/cores}
[junit4] 2> 49823 ERROR (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 49823 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 49824 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 49824 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 49824 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 49824 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T20:29:22.801Z
[junit4] 2> 49833 INFO (zkConnectionManagerCallback-98-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 49835 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
[junit4] 2> 49835 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/control-001/solr.xml
[junit4] 2> 49838 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
[junit4] 2> 49838 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 49845 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff, but no JMX reporters were configured - adding default JMX reporter.
[junit4] 2> 50047 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40523/solr
[junit4] 2> 50065 INFO (zkConnectionManagerCallback-102-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 50079 INFO (zkConnectionManagerCallback-104-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 50238 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 50240 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:39135_
[junit4] 2> 50241 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.c.Overseer Overseer (id=72235044170301445-127.0.0.1:39135_-n_0000000000) starting
[junit4] 2> 50286 INFO (zkConnectionManagerCallback-111-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 50289 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:40523/solr ready
[junit4] 2> 50399 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39135_
[junit4] 2> 50419 INFO (zkCallback-103-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 50422 INFO (zkCallback-110-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 50442 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 50457 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 50470 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 50471 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 50473 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:39135_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/control-001/cores
[junit4] 2> 50521 INFO (zkConnectionManagerCallback-117-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 50522 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 50524 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:40523/solr ready
[junit4] 2> 50554 INFO (qtp1135287928-1057) [n:127.0.0.1:39135_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:39135_&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 50572 INFO (OverseerThreadFactory-659-thread-1) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection control_collection
[junit4] 2> 50715 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 50716 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
[junit4] 2> 51858 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 51870 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
[junit4] 2> 52007 WARN (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Field lowerfilt1and2 is not multivalued and destination for multiple copyFields (2)
[junit4] 2> 52007 WARN (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Field text is not multivalued and destination for multiple copyFields (3)
[junit4] 2> 52008 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 52008 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' using configuration from collection control_collection, trusted=true
[junit4] 2> 52008 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.control_collection.shard1.replica_n1' (registry 'solr.core.control_collection.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 52008 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 52009 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/control-001/cores/control_collection_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/control-001/cores/control_collection_shard1_replica_n1/data/]
[junit4] 2> 52012 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=44, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.6011710914531239]
[junit4] 2> 52248 WARN (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 52329 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 52329 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 52331 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 52331 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 52332 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=45, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
[junit4] 2> 52333 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@723bee59[control_collection_shard1_replica_n1] main]
[junit4] 2> 52336 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 52336 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 52337 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 52339 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603824847752790016
[junit4] 2> 52340 INFO (searcherExecutor-664-thread-1-processing-n:127.0.0.1:39135_ x:control_collection_shard1_replica_n1 c:control_collection s:shard1) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [control_collection_shard1_replica_n1] Registered new searcher Searcher@723bee59[control_collection_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 52403 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, version=0}
[junit4] 2> 52417 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 52417 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 52417 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39135/control_collection_shard1_replica_n1/
[junit4] 2> 52418 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 52419 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:39135/control_collection_shard1_replica_n1/ has no replicas
[junit4] 2> 52419 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 52431 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39135/control_collection_shard1_replica_n1/ shard1
[junit4] 2> 52585 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 52629 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1916
[junit4] 2> 52644 INFO (qtp1135287928-1057) [n:127.0.0.1:39135_ ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
[junit4] 2> 52690 INFO (zkCallback-103-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1])
[junit4] 2> 53644 INFO (qtp1135287928-1057) [n:127.0.0.1:39135_ ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:39135_&wt=javabin&version=2} status=0 QTime=3091
[junit4] 2> 53653 INFO (zkConnectionManagerCallback-122-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 53654 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 53658 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:40523/solr ready
[junit4] 2> 53661 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false
[junit4] 2> 53665 INFO (qtp1135287928-1055) [n:127.0.0.1:39135_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=1&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 53692 INFO (OverseerCollectionConfigSetProcessor-72235044170301445-127.0.0.1:39135_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 53692 INFO (OverseerThreadFactory-659-thread-2) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1
[junit4] 2> 53693 WARN (OverseerThreadFactory-659-thread-2) [ ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to create a collection (collection1) without cores.
[junit4] 2> 53912 INFO (qtp1135287928-1055) [n:127.0.0.1:39135_ ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
[junit4] 2> 53912 INFO (qtp1135287928-1055) [n:127.0.0.1:39135_ ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=1&wt=javabin&version=2} status=0 QTime=247
[junit4] 2> 53999 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-1-001 of type NRT
[junit4] 2> 54002 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 54038 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 54038 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 54039 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 54043 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@599665a0{/,null,AVAILABLE}
[junit4] 2> 54043 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@2d53c46a{SSL,[ssl, http/1.1]}{127.0.0.1:43627}
[junit4] 2> 54043 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server Started @54074ms
[junit4] 2> 54044 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/tempDir-001/jetty1, solrconfig=solrconfig.xml, hostContext=/, hostPort=43627, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/../../../../../../../../../home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-1-001/cores}
[junit4] 2> 54044 ERROR (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 54044 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 54044 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 54044 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 54044 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 54044 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T20:29:27.021Z
[junit4] 2> 54078 INFO (zkConnectionManagerCallback-124-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 54082 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
[junit4] 2> 54082 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-1-001/solr.xml
[junit4] 2> 54086 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
[junit4] 2> 54087 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 54088 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff, but no JMX reporters were configured - adding default JMX reporter.
[junit4] 2> 54176 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40523/solr
[junit4] 2> 54182 INFO (zkConnectionManagerCallback-128-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 54187 INFO (zkConnectionManagerCallback-130-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 54195 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 54202 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 54209 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
[junit4] 2> 54209 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:43627_
[junit4] 2> 54212 INFO (zkCallback-103-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 54212 INFO (zkCallback-110-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 54212 INFO (zkCallback-121-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 54225 INFO (zkCallback-129-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 54235 INFO (zkConnectionManagerCallback-137-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 54237 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 54238 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:40523/solr ready
[junit4] 2> 54238 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 54252 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 54258 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 54258 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 54260 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:43627_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/../../../../../../../../../home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-1-001/cores
[junit4] 2> 54278 INFO (qtp1135287928-1057) [n:127.0.0.1:39135_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:43627_&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 54281 INFO (OverseerCollectionConfigSetProcessor-72235044170301445-127.0.0.1:39135_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 54289 INFO (OverseerThreadFactory-659-thread-3) [ c:collection1 s:shard2 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:43627_ for creating new replica
[junit4] 2> 54328 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ x:collection1_shard2_replica_n41] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 55342 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 55355 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema [collection1_shard2_replica_n41] Schema name=test
[junit4] 2> 55414 WARN (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema Field lowerfilt1and2 is not multivalued and destination for multiple copyFields (2)
[junit4] 2> 55414 WARN (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema Field text is not multivalued and destination for multiple copyFields (3)
[junit4] 2> 55414 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 55415 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard2_replica_n41' using configuration from collection collection1, trusted=true
[junit4] 2> 55415 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard2.replica_n41' (registry 'solr.core.collection1.shard2.replica_n41') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 55415 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 55415 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SolrCore [[collection1_shard2_replica_n41] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-1-001/cores/collection1_shard2_replica_n41], dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/../../../../../../../../../home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-1-001/cores/collection1_shard2_replica_n41/data/]
[junit4] 2> 55418 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=44, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.6011710914531239]
[junit4] 2> 55428 WARN (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 55473 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 55473 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 55474 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 55474 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 55476 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=45, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
[junit4] 2> 55476 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.SolrIndexSearcher Opening [Searcher@6ca02341[collection1_shard2_replica_n41] main]
[junit4] 2> 55479 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 55479 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 55480 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 55480 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603824851046367232
[junit4] 2> 55483 INFO (searcherExecutor-678-thread-1-processing-n:127.0.0.1:43627_ x:collection1_shard2_replica_n41 c:collection1 s:shard2) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SolrCore [collection1_shard2_replica_n41] Registered new searcher Searcher@6ca02341[collection1_shard2_replica_n41] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 55511 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={core_node42=0}, version=0}
[junit4] 2> 55518 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 55518 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 55518 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43627/collection1_shard2_replica_n41/
[junit4] 2> 55518 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 55518 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy https://127.0.0.1:43627/collection1_shard2_replica_n41/ has no replicas
[junit4] 2> 55518 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 55522 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43627/collection1_shard2_replica_n41/ shard2
[junit4] 2> 55676 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 55679 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1351
[junit4] 2> 55681 INFO (qtp1135287928-1057) [n:127.0.0.1:39135_ c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:43627_&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2} status=0 QTime=1403
[junit4] 2> 55761 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-2-001 of type NRT
[junit4] 2> 55762 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 55767 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 55767 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 55767 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 55767 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@48fa3fab{/,null,AVAILABLE}
[junit4] 2> 55768 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@41fc6f9d{SSL,[ssl, http/1.1]}{127.0.0.1:44675}
[junit4] 2> 55768 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server Started @55799ms
[junit4] 2> 55768 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/tempDir-001/jetty2, solrconfig=solrconfig.xml, hostContext=/, hostPort=44675, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-2-001/cores}
[junit4] 2> 55768 ERROR (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 55769 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 55769 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 55769 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 55769 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 55769 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T20:29:28.746Z
[junit4] 2> 55771 INFO (zkConnectionManagerCallback-140-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 55774 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
[junit4] 2> 55774 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-2-001/solr.xml
[junit4] 2> 55777 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
[junit4] 2> 55777 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 55778 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff, but no JMX reporters were configured - adding default JMX reporter.
[junit4] 2> 55900 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40523/solr
[junit4] 2> 55938 INFO (zkConnectionManagerCallback-144-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 55948 INFO (zkConnectionManagerCallback-146-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 55953 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 55955 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 55958 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
[junit4] 2> 55958 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:44675_
[junit4] 2> 55959 INFO (zkCallback-136-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 55959 INFO (zkCallback-110-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 55959 INFO (zkCallback-129-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 55959 INFO (zkCallback-121-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 55963 INFO (zkCallback-103-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 55975 INFO (zkCallback-145-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 55994 INFO (zkConnectionManagerCallback-153-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 55995 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 55996 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:40523/solr ready
[junit4] 2> 55996 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 56011 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 56019 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 56019 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 56020 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:44675_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-2-001/cores
[junit4] 2> 56072 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:44675_&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 56079 INFO (OverseerCollectionConfigSetProcessor-72235044170301445-127.0.0.1:39135_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000004 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 56083 INFO (OverseerThreadFactory-659-thread-4) [ c:collection1 s:shard1 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:44675_ for creating new replica
[junit4] 2> 56106 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ x:collection1_shard1_replica_n43] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n43&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 57119 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 57157 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.s.IndexSchema [collection1_shard1_replica_n43] Schema name=test
[junit4] 2> 57230 WARN (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.s.IndexSchema Field lowerfilt1and2 is not multivalued and destination for multiple copyFields (2)
[junit4] 2> 57230 WARN (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.s.IndexSchema Field text is not multivalued and destination for multiple copyFields (3)
[junit4] 2> 57231 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 57231 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_n43' using configuration from collection collection1, trusted=true
[junit4] 2> 57231 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard1.replica_n43' (registry 'solr.core.collection1.shard1.replica_n43') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 57231 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 57231 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.SolrCore [[collection1_shard1_replica_n43] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-2-001/cores/collection1_shard1_replica_n43], dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-2-001/cores/collection1_shard1_replica_n43/data/]
[junit4] 2> 57234 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=44, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.6011710914531239]
[junit4] 2> 57235 WARN (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 57282 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 57282 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 57283 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 57283 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 57284 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=45, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
[junit4] 2> 57285 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.s.SolrIndexSearcher Opening [Searcher@7aece9b6[collection1_shard1_replica_n43] main]
[junit4] 2> 57286 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 57286 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 57287 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 57287 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603824852941144064
[junit4] 2> 57289 INFO (searcherExecutor-692-thread-1-processing-n:127.0.0.1:44675_ x:collection1_shard1_replica_n43 c:collection1 s:shard1) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.SolrCore [collection1_shard1_replica_n43] Registered new searcher Searcher@7aece9b6[collection1_shard1_replica_n43] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 57306 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node44=0}, version=0}
[junit4] 2> 57308 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 57308 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 57308 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:44675/collection1_shard1_replica_n43/
[junit4] 2> 57308 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 57308 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.SyncStrategy https://127.0.0.1:44675/collection1_shard1_replica_n43/ has no replicas
[junit4] 2> 57308 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 57315 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:44675/collection1_shard1_replica_n43/ shard1
[junit4] 2> 57469 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 57472 INFO (qtp98666890-1156) [n:127.0.0.1:44675_ c:collection1 s:shard1 x:collection1_shard1_replica_n43] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n43&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1365
[junit4] 2> 57474 INFO (qtp1135287928-1053) [n:127.0.0.1:39135_ c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:44675_&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2} status=0 QTime=1402
[junit4] 2> 57550 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 3 in directory /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-3-001 of type NRT
[junit4] 2> 57552 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
[junit4] 2> 57554 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 57554 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 57554 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 57554 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@546c5f1e{/,null,AVAILABLE}
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@7d8dbab8{SSL,[ssl, http/1.1]}{127.0.0.1:40885}
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.e.j.s.Server Started @57585ms
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/tempDir-001/jetty3, solrconfig=solrconfig.xml, hostContext=/, hostPort=40885, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-3-001/cores}
[junit4] 2> 57555 ERROR (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 57555 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-20T20:29:30.532Z
[junit4] 2> 57558 INFO (zkConnectionManagerCallback-156-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 57560 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
[junit4] 2> 57560 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-3-001/solr.xml
[junit4] 2> 57562 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
[junit4] 2> 57562 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 57563 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff, but no JMX reporters were configured - adding default JMX reporter.
[junit4] 2> 57638 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40523/solr
[junit4] 2> 57641 INFO (zkConnectionManagerCallback-160-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 57646 INFO (zkConnectionManagerCallback-162-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 57654 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 57656 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 57657 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
[junit4] 2> 57657 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:40885_
[junit4] 2> 57658 INFO (zkCallback-136-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57658 INFO (zkCallback-129-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57658 INFO (zkCallback-152-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57658 INFO (zkCallback-145-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57658 INFO (zkCallback-110-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57659 INFO (zkCallback-121-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57659 INFO (zkCallback-103-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57659 INFO (zkCallback-161-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 57679 INFO (zkConnectionManagerCallback-169-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 57681 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
[junit4] 2> 57682 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:40523/solr ready
[junit4] 2> 57682 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 57707 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 57715 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 57715 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 57716 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [n:127.0.0.1:40885_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-3-001/cores
[junit4] 2> 57752 INFO (qtp1135287928-1057) [n:127.0.0.1:39135_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:40885_&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 57757 INFO (OverseerCollectionConfigSetProcessor-72235044170301445-127.0.0.1:39135_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000006 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 57758 INFO (OverseerThreadFactory-659-thread-5) [ c:collection1 s:shard2 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:40885_ for creating new replica
[junit4] 2> 57777 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ x:collection1_shard2_replica_n45] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n45&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 58792 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
[junit4] 2> 58830 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.s.IndexSchema [collection1_shard2_replica_n45] Schema name=test
[junit4] 2> 58896 WARN (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.s.IndexSchema Field lowerfilt1and2 is not multivalued and destination for multiple copyFields (2)
[junit4] 2> 58896 WARN (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.s.IndexSchema Field text is not multivalued and destination for multiple copyFields (3)
[junit4] 2> 58896 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 58896 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard2_replica_n45' using configuration from collection collection1, trusted=true
[junit4] 2> 58897 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard2.replica_n45' (registry 'solr.core.collection1.shard2.replica_n45') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4f34ceff
[junit4] 2> 58897 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 58897 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.SolrCore [[collection1_shard2_replica_n45] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-3-001/cores/collection1_shard2_replica_n45], dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp/solr.update.processor.TestNamedUpdateProcessors_AB5C7C8C5A9F386C-001/shard-3-001/cores/collection1_shard2_replica_n45/data/]
[junit4] 2> 58899 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=44, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.6011710914531239]
[junit4] 2> 58915 WARN (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 58953 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 58953 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 58976 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 58976 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 58977 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=45, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
[junit4] 2> 58979 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.s.SolrIndexSearcher Opening [Searcher@6f930454[collection1_shard2_replica_n45] main]
[junit4] 2> 58983 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 58983 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 58984 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 58985 INFO (searcherExecutor-706-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.SolrCore [collection1_shard2_replica_n45] Registered new searcher Searcher@6f930454[collection1_shard2_replica_n45] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 58987 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603824854723723264
[junit4] 2> 58995 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={core_node42=0, core_node46=0}, version=1}
[junit4] 2> 58996 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.c.ZkController Core needs to recover:collection1_shard2_replica_n45
[junit4] 2> 59007 INFO (updateExecutor-157-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.u.DefaultSolrCoreState Running recovery
[junit4] 2> 59013 INFO (recoveryExecutor-158-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy Starting recovery process. recoveringAfterStartup=true
[junit4] 2> 59014 INFO (recoveryExecutor-158-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy ###### startupVersions=[[]]
[junit4] 2> 59075 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 r:core_node42 x:collection1_shard2_replica_n41] o.a.s.c.S.Request [collection1_shard2_replica_n41] webapp= path=/admin/ping params={wt=javabin&version=2} hits=0 status=0 QTime=2
[junit4] 2> 59076 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ c:collection1 s:shard2 r:core_node42 x:collection1_shard2_replica_n41] o.a.s.c.S.Request [collection1_shard2_replica_n41] webapp= path=/admin/ping params={wt=javabin&version=2} status=0 QTime=2
[junit4] 2> 59076 INFO (recoveryExecutor-158-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy Begin buffering updates. core=[collection1_shard2_replica_n45]
[junit4] 2> 59076 INFO (recoveryExecutor-158-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
[junit4] 2> 59076 INFO (recoveryExecutor-158-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy Publishing state of core [collection1_shard2_replica_n45] as recovering, leader is [https://127.0.0.1:43627/collection1_shard2_replica_n41/] and I am [https://127.0.0.1:40885/collection1_shard2_replica_n45/]
[junit4] 2> 59093 INFO (qtp1874102241-1193) [n:127.0.0.1:40885_ c:collection1 s:shard2 x:collection1_shard2_replica_n45] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n45&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1316
[junit4] 2> 59096 INFO (recoveryExecutor-158-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={core_node42=0, core_node46=0, core_node46_recovering=0}, version=2}
[junit4] 2> 59105 INFO (qtp1135287928-1057) [n:127.0.0.1:39135_ c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:40885_&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2} status=0 QTime=1353
[junit4] 2> 59121 INFO (recoveryExecutor-158-thread-1-processing-n:127.0.0.1:40885_ x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:40885_ c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy Sending prep recovery command to [https://127.0.0.1:43627]; [WaitForState: action=PREPRECOVERY&core=collection1_shard2_replica_n41&nodeName=127.0.0.1:40885_&coreNodeName=core_node46&state=recovering&checkLive=true&onlyIfLeader=true&onlyIfLeaderActive=true]
[junit4] 2> 59150 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ x:collection1_shard2_replica_n41] o.a.s.h.a.PrepRecoveryOp Going to wait for coreNodeName: core_node46, state: recovering, checkLive: true, onlyIfLeader: true, onlyIfLeaderActive: true, maxTime: 183 s
[junit4] 2> 59152 INFO (qtp673533604-1117) [n:127.0.0.1:43627_ x:collection1_shard2_replica_n41] o.a.s.h.a.PrepRecoveryOp In WaitForState(recovering): collection=collection1, shard=shard2, thisCore=collection1_shard2_replica_n41, leaderDoesNotNeedRecovery=false, isLeader? true, live=true, checkLive=true, currentState=down, localState=active, nodeName=127.0.0.1:40885_, coreNodeName=core_node46, onlyIfActiveCheckResult=false, nodeProps: core_node46:{"core":"collection1_shard2_replica_n45","base_url":"https://127.0.0.1:40885","node_name":"127.0.0.1:40885_","state":"down","type":"NRT"}
[junit4] 2> 59241 INFO (TEST-TestNamedUpdateProcessors.test-seed#[AB5C7C8C5A9F386C]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 4 in directory /home/jenkins/worksp
[...truncated too long message...]
solve:
init:
-clover.disable:
-clover.load:
-clover.classpath:
-clover.setup:
clover:
compile-core:
init:
-clover.disable:
-clover.load:
-clover.classpath:
-clover.setup:
clover:
common.compile-core:
common-solr.compile-core:
compile-core:
compile-test-framework:
-check-git-state:
-git-cleanroot:
-copy-git-state:
git-autoclean:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file = /home/jenkins/workspace/Lucene-Solr-7.x-Linux/lucene/top-level-ivy-settings.xml
resolve:
init:
compile-lucene-core:
compile-codecs:
-check-git-state:
-git-cleanroot:
-copy-git-state:
git-autoclean:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file = /home/jenkins/workspace/Lucene-Solr-7.x-Linux/lucene/top-level-ivy-settings.xml
resolve:
common.init:
compile-lucene-core:
init:
-clover.disable:
-clover.load:
-clover.classpath:
-clover.setup:
clover:
compile-core:
-clover.disable:
-clover.load:
-clover.classpath:
-clover.setup:
clover:
common.compile-core:
compile-core:
common.compile-test:
[mkdir] Created dir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/classes/test
[javac] Compiling 917 source files to /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/classes/test
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.
[javac] Creating empty /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/classes/test/org/apache/solr/cloud/autoscaling/sim/package-info.class
[copy] Copying 1 file to /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/classes/test
common-solr.compile-test:
compile-test:
BUILD SUCCESSFUL
Total time: 56 seconds
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.TestNamedUpdateProcessors" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseParallelGC" -Dtests.seed=AB5C7C8C5A9F386C -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=he-IL -Dtests.timezone=Atlantic/St_Helena -Dtests.asserts=true -Dtests.file.encoding=UTF-8
Buildfile: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/build.xml
-clover.disable:
ivy-configure:
[ivy:configure] :: Apache Ivy 2.4.0 - 20141213170938 :: http://ant.apache.org/ivy/ ::
[ivy:configure] :: loading settings :: file = /home/jenkins/workspace/Lucene-Solr-7.x-Linux/lucene/top-level-ivy-settings.xml
install-junit4-taskdef:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
resolve-groovy:
[ivy:cachepath] :: resolving dependencies :: org.codehaus.groovy#groovy-all-caller;working
[ivy:cachepath] confs: [default]
[ivy:cachepath] found org.codehaus.groovy#groovy-all;2.4.15 in public
[ivy:cachepath] :: resolution report :: resolve 50ms :: artifacts dl 1ms
---------------------------------------------------------------------
| | modules || artifacts |
| conf | number| search|dwnlded|evicted|| number|dwnlded|
---------------------------------------------------------------------
| default | 1 | 0 | 0 | 0 || 1 | 0 |
---------------------------------------------------------------------
-init-totals:
-test:
[mkdir] Created dir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test
[junit4:pickseed] Seed property 'tests.seed' already defined: AB5C7C8C5A9F386C
[mkdir] Created dir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/temp
[junit4] <JUnit4> says नमस्ते! Master seed: AB5C7C8C5A9F386C
[junit4] Executing 5 suites with 3 JVMs.
[junit4]
[junit4] Started J0 PID(13137@serv1.sd-datasolutions.de).
[junit4] Started J2 PID(13136@serv1.sd-datasolutions.de).
[junit4] Started J1 PID(13138@serv1.sd-datasolutions.de).
[junit4] Suite: org.apache.solr.update.processor.TestNamedUpdateProcessors
[junit4] OK 28.5s J2 | TestNamedUpdateProcessors.test
[junit4] Completed [1/5] on J2 in 30.64s, 1 test
[junit4]
[junit4] Suite: org.apache.solr.update.processor.TestNamedUpdateProcessors
[junit4] OK 28.1s J0 | TestNamedUpdateProcessors.test
[junit4] Completed [2/5] on J0 in 30.71s, 1 test
[junit4]
[junit4] Duplicate suite name used with XML reports: org.apache.solr.update.processor.TestNamedUpdateProcessors. This may confuse tools that process XML reports. Set 'ignoreDuplicateSuites' to true to skip this message.
[junit4] Suite: org.apache.solr.update.processor.TestNamedUpdateProcessors
[junit4] OK 30.4s J1 | TestNamedUpdateProcessors.test
[junit4] Completed [3/5] on J1 in 32.70s, 1 test
[junit4]
[junit4] Suite: org.apache.solr.update.processor.TestNamedUpdateProcessors
[junit4] OK 18.2s J0 | TestNamedUpdateProcessors.test
[junit4] Completed [4/5] on J0 in 18.85s, 1 test
[junit4]
[junit4] Suite: org.apache.solr.update.processor.TestNamedUpdateProcessors
[junit4] OK 19.3s J2 | TestNamedUpdateProcessors.test
[junit4] Completed [5/5] on J2 in 19.47s, 1 test
[junit4]
[junit4] JVM J0: 0.42 .. 51.39 = 50.97s
[junit4] JVM J1: 0.43 .. 34.07 = 33.64s
[junit4] JVM J2: 0.42 .. 51.68 = 51.25s
[junit4] Execution time total: 51 seconds
[junit4] Tests summary: 5 suites, 5 tests
[junit4] Could not remove temporary path: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2 (java.nio.file.DirectoryNotEmptyException: Remaining files: [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp])
[junit4] Could not remove temporary path: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0 (java.nio.file.DirectoryNotEmptyException: Remaining files: [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J0/temp])
[junit4] Could not remove temporary path: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1 (java.nio.file.DirectoryNotEmptyException: Remaining files: [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp])
[echo] 5 slowest tests:
[junit4:tophints] 284.89s | org.apache.solr.cloud.api.collections.ShardSplitTest
[junit4:tophints] 116.72s | org.apache.solr.update.SoftAutoCommitTest
[junit4:tophints] 109.96s | org.apache.solr.cloud.ZkFailoverTest
[junit4:tophints] 106.08s | org.apache.solr.cloud.TestTlogReplica
[junit4:tophints] 100.33s | org.apache.solr.cloud.autoscaling.AutoAddReplicasIntegrationTest
-check-totals:
test-nocompile:
BUILD SUCCESSFUL
Total time: 53 seconds
[repro] Failures:
[repro] 0/5 failed: org.apache.solr.update.processor.TestNamedUpdateProcessors
[repro] Exiting with code 0
+ mv lucene/build lucene/build.repro
+ mv solr/build solr/build.repro
+ mv lucene/build.orig lucene/build
+ mv solr/build.orig solr/build
Archiving artifacts
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
java.lang.InterruptedException: no matches found within 10000
at hudson.FilePath$ValidateAntFileMask.hasMatch(FilePath.java:2843)
at hudson.FilePath$ValidateAntFileMask.invoke(FilePath.java:2722)
at hudson.FilePath$ValidateAntFileMask.invoke(FilePath.java:2703)
at hudson.FilePath.act(FilePath.java:1095)
at hudson.FilePath.act(FilePath.java:1078)
at hudson.FilePath.validateAntFileMask(FilePath.java:2701)
at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
at hudson.model.Build$BuildExecution.post2(Build.java:186)
at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
at hudson.model.Run.execute(Run.java:1823)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:97)
at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern "**/*.events,heapdumps/**,**/*_pid*.log". Configuration error?
[WARNINGS] Parsing warnings in console log with parser Java Compiler (javac)
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=3b0edb0d667dbfa8c8ffb6c836a68a6f07effc00, workspace=/var/lib/jenkins/workspace/Lucene-Solr-7.x-Linux
[WARNINGS] Computing warning deltas based on reference build #2162
Recording test results
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2