You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2018/12/23 08:44:32 UTC
[JENKINS] Lucene-Solr-7.6-Linux (32bit/jdk1.8.0_172) - Build # 139
- Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.6-Linux/139/
Java: 32bit/jdk1.8.0_172 -client -XX:+UseParallelGC
1 tests failed.
FAILED: org.apache.solr.cloud.MoveReplicaHDFSFailoverTest.testOldReplicaIsDeletedInRaceCondition
Error Message:
Error from server at https://127.0.0.1:38621/solr: Could not fully remove collection: movereplicatest_coll4
Stack Trace:
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at https://127.0.0.1:38621/solr: Could not fully remove collection: movereplicatest_coll4
at __randomizedtesting.SeedInfo.seed([4270250C0D0814BA:4820AA7ACD08751B]:0)
at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483)
at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:413)
at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1107)
at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:884)
at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:211)
at org.apache.solr.cloud.MoveReplicaHDFSFailoverTest.testOldReplicaIsDeletedInRaceCondition(MoveReplicaHDFSFailoverTest.java:195)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
Build Log:
[...truncated 14118 lines...]
[junit4] Suite: org.apache.solr.cloud.MoveReplicaHDFSFailoverTest
[junit4] 2> 1560371 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/init-core-data-001
[junit4] 2> 1560372 WARN (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=6 numCloses=6
[junit4] 2> 1560372 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 1560374 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
[junit4] 2> 1560375 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in /home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-001
[junit4] 2> 1560375 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1560375 INFO (Thread-3837) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1560375 INFO (Thread-3837) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 1560377 ERROR (Thread-3837) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 1560475 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.c.ZkTestServer start zk server on port:35221
[junit4] 2> 1560479 INFO (zkConnectionManagerCallback-3498-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1560486 INFO (jetty-launcher-3495-thread-1) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
[junit4] 2> 1560487 INFO (jetty-launcher-3495-thread-2) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-1) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-2) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-2) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-2) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@189d60{/solr,null,AVAILABLE}
[junit4] 2> 1560508 INFO (jetty-launcher-3495-thread-2) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@ae680d{/solr,null,AVAILABLE}
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@189d02a{SSL,[ssl, http/1.1]}{127.0.0.1:46313}
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-2) [ ] o.e.j.s.AbstractConnector Started ServerConnector@8ebfbe{SSL,[ssl, http/1.1]}{127.0.0.1:38621}
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-1) [ ] o.e.j.s.Server Started @1560538ms
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-2) [ ] o.e.j.s.Server Started @1560538ms
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=46313}
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=38621}
[junit4] 2> 1560510 ERROR (jetty-launcher-3495-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1560510 ERROR (jetty-launcher-3495-thread-2) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.6.0
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.6.0
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1560510 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1560511 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-12-23T08:08:25.227Z
[junit4] 2> 1560511 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-12-23T08:08:25.228Z
[junit4] 2> 1560517 INFO (zkConnectionManagerCallback-3501-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1560518 INFO (zkConnectionManagerCallback-3502-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1560518 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1560518 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1560527 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x1001d31808d0001, likely client has closed socket
[junit4] 2> 1560913 INFO (jetty-launcher-3495-thread-2) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:35221/solr
[junit4] 2> 1560914 INFO (zkConnectionManagerCallback-3506-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1560916 INFO (zkConnectionManagerCallback-3508-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1560979 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:38621_solr
[junit4] 2> 1560980 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.c.Overseer Overseer (id=72089692485255172-127.0.0.1:38621_solr-n_0000000000) starting
[junit4] 2> 1560984 INFO (zkConnectionManagerCallback-3515-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1560985 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:35221/solr ready
[junit4] 2> 1560986 INFO (OverseerStateUpdate-72089692485255172-127.0.0.1:38621_solr-n_0000000000) [n:127.0.0.1:38621_solr ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:38621_solr
[junit4] 2> 1560988 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:38621_solr
[junit4] 2> 1561001 INFO (zkCallback-3507-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1561002 INFO (zkCallback-3514-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1561006 DEBUG (OverseerAutoScalingTriggerThread-72089692485255172-127.0.0.1:38621_solr-n_0000000000) [ ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:38621_solr]
[junit4] 2> 1561012 DEBUG (ScheduledTrigger-6910-thread-1) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
[junit4] 2> 1561019 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1561044 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38621.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1561052 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38621.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1561053 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38621.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1561054 INFO (jetty-launcher-3495-thread-2) [n:127.0.0.1:38621_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-001/node2/.
[junit4] 2> 1562014 DEBUG (ScheduledTrigger-6910-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
[junit4] 2> 1562289 INFO (jetty-launcher-3495-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:35221/solr
[junit4] 2> 1562290 INFO (zkConnectionManagerCallback-3520-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562293 INFO (zkConnectionManagerCallback-3522-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562299 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1562302 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1562302 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:46313_solr
[junit4] 2> 1562303 INFO (zkCallback-3507-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1562304 INFO (zkCallback-3521-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1562304 INFO (zkCallback-3514-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1562330 INFO (zkConnectionManagerCallback-3529-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562331 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1562332 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:35221/solr ready
[junit4] 2> 1562334 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1562350 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46313.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1562360 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46313.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1562360 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46313.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1562362 INFO (jetty-launcher-3495-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-001/node1/.
[junit4] 2> 1562390 INFO (zkConnectionManagerCallback-3532-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562395 INFO (zkConnectionManagerCallback-3537-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1562396 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1562397 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:35221/solr ready
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 1562512 WARN (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 1562521 WARN (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1562523 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1562536 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_localdomain_46081_hdfs____32mp8h/webapp
[junit4] 2> 1563014 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:46081
[junit4] 2> 1563016 DEBUG (ScheduledTrigger-6910-thread-4) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
[junit4] 2> 1563126 WARN (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1563127 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1563139 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_34621_datanode____dz0pgb/webapp
[junit4] 2> 1563582 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34621
[junit4] 2> 1563634 WARN (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1563635 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1563647 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_45313_datanode____msur17/webapp
[junit4] 2> 1563780 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:43569) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1563796 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x1bd75639724c8: from storage DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1 node DatanodeRegistration(127.0.0.1:46851, datanodeUuid=8ff5a187-be2d-435b-810c-df8b67206c6f, infoPort=41063, infoSecurePort=0, ipcPort=45873, storageInfo=lv=-56;cid=testClusterID;nsid=1121442574;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 1563796 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x1bd75639724c8: from storage DS-35bb0c57-eb2d-4410-9619-0e444cca27f4 node DatanodeRegistration(127.0.0.1:46851, datanodeUuid=8ff5a187-be2d-435b-810c-df8b67206c6f, infoPort=41063, infoSecurePort=0, ipcPort=45873, storageInfo=lv=-56;cid=testClusterID;nsid=1121442574;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1564016 DEBUG (ScheduledTrigger-6910-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
[junit4] 2> 1564179 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:45313
[junit4] 2> 1564470 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:43569) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1564478 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x1bd758c52abdb: from storage DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2 node DatanodeRegistration(127.0.0.1:37925, datanodeUuid=ec6c3b42-f20b-4412-9bd7-75c50b5af405, infoPort=45985, infoSecurePort=0, ipcPort=40607, storageInfo=lv=-56;cid=testClusterID;nsid=1121442574;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 1564479 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0x1bd758c52abdb: from storage DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51 node DatanodeRegistration(127.0.0.1:37925, datanodeUuid=ec6c3b42-f20b-4412-9bd7-75c50b5af405, infoPort=45985, infoSecurePort=0, ipcPort=40607, storageInfo=lv=-56;cid=testClusterID;nsid=1121442574;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1564692 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.SolrTestCaseJ4 ###Starting testOldReplicaIsDeleted
[junit4] 2> 1564739 INFO (qtp19430863-15894) [n:127.0.0.1:38621_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=movereplicatest_coll3&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:38621_solr&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1564743 INFO (OverseerThreadFactory-6912-thread-1-processing-n:127.0.0.1:38621_solr) [n:127.0.0.1:38621_solr ] o.a.s.c.a.c.CreateCollectionCmd Create collection movereplicatest_coll3
[junit4] 2> 1564861 INFO (qtp19430863-16148) [n:127.0.0.1:38621_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=1
[junit4] 2> 1564909 INFO (qtp32863764-15887) [n:127.0.0.1:46313_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=1
[junit4] 2> 1564915 INFO (OverseerStateUpdate-72089692485255172-127.0.0.1:38621_solr-n_0000000000) [n:127.0.0.1:38621_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"movereplicatest_coll3",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"movereplicatest_coll3_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:38621/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1565017 DEBUG (ScheduledTrigger-6910-thread-4) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
[junit4] 2> 1565122 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr x:movereplicatest_coll3_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node2&collection.configName=conf1&newCollection=true&name=movereplicatest_coll3_shard1_replica_n1&action=CREATE&numShards=1&collection=movereplicatest_coll3&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1565122 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1566017 DEBUG (ScheduledTrigger-6910-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
[junit4] 2> 1566142 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1566153 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.IndexSchema [movereplicatest_coll3_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1566157 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1566157 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'movereplicatest_coll3_shard1_replica_n1' using configuration from collection movereplicatest_coll3, trusted=true
[junit4] 2> 1566157 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38621.solr.core.movereplicatest_coll3.shard1.replica_n1' (registry 'solr.core.movereplicatest_coll3.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1566164 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:43569/data
[junit4] 2> 1566164 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1566164 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrCore [[movereplicatest_coll3_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-001/node2/movereplicatest_coll3_shard1_replica_n1], dataDir=[hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/]
[junit4] 2> 1566167 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/snapshot_metadata
[junit4] 2> 1566205 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data
[junit4] 2> 1566230 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/index
[junit4] 2> 1566294 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|FINALIZED]]} size 0
[junit4] 2> 1566299 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741825_1001 size 69
[junit4] 2> 1566379 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1566379 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1566379 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1566394 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1566394 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1566460 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@8d4d42[movereplicatest_coll3_shard1_replica_n1] main]
[junit4] 2> 1566462 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1566462 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1566464 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1566467 WARN (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1566468 INFO (searcherExecutor-6924-thread-1-processing-n:127.0.0.1:38621_solr x:movereplicatest_coll3_shard1_replica_n1 c:movereplicatest_coll3 s:shard1 r:core_node2) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrCore [movereplicatest_coll3_shard1_replica_n1] Registered new searcher Searcher@8d4d42[movereplicatest_coll3_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1566468 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1620629269968322560
[junit4] 2> 1566474 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/movereplicatest_coll3/terms/shard1 to Terms{values={core_node2=0}, version=0}
[junit4] 2> 1566476 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1566476 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1566476 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:38621/solr/movereplicatest_coll3_shard1_replica_n1/
[junit4] 2> 1566477 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1566477 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:38621/solr/movereplicatest_coll3_shard1_replica_n1/ has no replicas
[junit4] 2> 1566477 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1566480 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:38621/solr/movereplicatest_coll3_shard1_replica_n1/ shard1
[junit4] 2> 1566582 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1566585 INFO (qtp19430863-15888) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node2&collection.configName=conf1&newCollection=true&name=movereplicatest_coll3_shard1_replica_n1&action=CREATE&numShards=1&collection=movereplicatest_coll3&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1463
[junit4] 2> 1566589 INFO (qtp19430863-15894) [n:127.0.0.1:38621_solr ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
[junit4] 2> 1566684 INFO (zkCallback-3507-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/movereplicatest_coll3/state.json] for collection [movereplicatest_coll3] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1566744 INFO (OverseerCollectionConfigSetProcessor-72089692485255172-127.0.0.1:38621_solr-n_0000000000) [n:127.0.0.1:38621_solr ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 1567017 DEBUG (ScheduledTrigger-6910-thread-4) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
[junit4] 2> 1567590 INFO (qtp19430863-15894) [n:127.0.0.1:38621_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=movereplicatest_coll3&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:38621_solr&wt=javabin&version=2} status=0 QTime=2851
[junit4] 2> 1567609 INFO (qtp19430863-15893) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/movereplicatest_coll3/terms/shard1 to Terms{values={core_node2=1}, version=1}
[junit4] 2> 1567633 INFO (qtp19430863-15893) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [movereplicatest_coll3_shard1_replica_n1] webapp=/solr path=/update params={wt=javabin&version=2}{add=[1 (1620629271149019136)]} 0 40
[junit4] 2> 1567640 INFO (qtp19430863-15894) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [movereplicatest_coll3_shard1_replica_n1] webapp=/solr path=/update params={wt=javabin&version=2}{add=[2 (1620629271193059328)]} 0 4
[junit4] 2> 1567643 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.e.j.s.AbstractConnector Stopped ServerConnector@8ebfbe{SSL,[ssl, http/1.1]}{127.0.0.1:0}
[junit4] 2> 1567644 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.c.CoreContainer Shutting down CoreContainer instance=31056095
[junit4] 2> 1567645 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.node, tag=null
[junit4] 2> 1567645 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@1b8d626: rootName = solr_38621, domain = solr.node, service url = null, agent id = null] for registry solr.node / com.codahale.metrics.MetricRegistry@1d53c0d
[junit4] 2> 1567654 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.jvm, tag=null
[junit4] 2> 1567654 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@5619bb: rootName = solr_38621, domain = solr.jvm, service url = null, agent id = null] for registry solr.jvm / com.codahale.metrics.MetricRegistry@b0f23c
[junit4] 2> 1567659 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.jetty, tag=null
[junit4] 2> 1567660 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@184a4a9: rootName = solr_38621, domain = solr.jetty, service url = null, agent id = null] for registry solr.jetty / com.codahale.metrics.MetricRegistry@1d029e2
[junit4] 2> 1567662 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.c.ZkController Remove node as live in ZooKeeper:/live_nodes/127.0.0.1:38621_solr
[junit4] 2> 1567662 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.cluster, tag=null
[junit4] 2> 1567663 INFO (zkCallback-3521-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
[junit4] 2> 1567663 INFO (zkCallback-3514-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
[junit4] 2> 1567663 INFO (zkCallback-3528-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
[junit4] 2> 1567663 INFO (zkCallback-3507-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
[junit4] 2> 1567663 INFO (zkCallback-3536-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
[junit4] 2> 1567666 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrCore [movereplicatest_coll3_shard1_replica_n1] CLOSING SolrCore org.apache.solr.core.SolrCore@1688f58
[junit4] 2> 1567666 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.movereplicatest_coll3.shard1.replica_n1, tag=1688f58
[junit4] 2> 1567666 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@175f0f2: rootName = solr_38621, domain = solr.core.movereplicatest_coll3.shard1.replica_n1, service url = null, agent id = null] for registry solr.core.movereplicatest_coll3.shard1.replica_n1 / com.codahale.metrics.MetricRegistry@122c21c
[junit4] 2> 1567675 WARN (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1567676 WARN (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1567677 WARN (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1567684 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.movereplicatest_coll3.shard1.leader, tag=1688f58
[junit4] 2> 1567684 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 Committing on IndexWriter close.
[junit4] 2> 1567684 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@1932ed7 commitCommandVersion:0
[junit4] 2> 1567707 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741827_1003{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567709 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741827_1003 size 186
[junit4] 2> 1567734 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567736 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567759 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741829_1005{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567760 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741829_1005 size 59
[junit4] 2> 1567765 INFO (zkCallback-3507-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/movereplicatest_coll3/state.json] for collection [movereplicatest_coll3] has occurred - updating... (live nodes size: [1])
[junit4] 2> 1567781 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741830_1006{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567782 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741830_1006 size 83
[junit4] 2> 1567804 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741831_1007{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567805 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741831_1007{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567827 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741832_1008{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567828 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741832_1008{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567845 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741833_1009{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|FINALIZED]]} size 0
[junit4] 2> 1567845 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741833_1009{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|FINALIZED], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|FINALIZED]]} size 0
[junit4] 2> 1567862 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567863 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|FINALIZED]]} size 0
[junit4] 2> 1567881 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741835_1011{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567882 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741835_1011{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567901 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741836_1012{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|FINALIZED]]} size 0
[junit4] 2> 1567902 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741836_1012{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|FINALIZED], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|FINALIZED]]} size 0
[junit4] 2> 1567922 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741837_1013{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567923 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741837_1013 size 179
[junit4] 2> 1567942 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741838_1014{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567942 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741838_1014{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|FINALIZED]]} size 0
[junit4] 2> 1567960 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741839_1015{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567960 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741839_1015{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e835ce24-a3ca-4dad-a95b-ccbcf422dfe2:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567973 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741840_1016{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1567974 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741840_1016{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|FINALIZED]]} size 0
[junit4] 2> 1567989 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741841_1017{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-35bb0c57-eb2d-4410-9619-0e444cca27f4:NORMAL:127.0.0.1:46851|RBW]]} size 0
[junit4] 2> 1567990 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741841_1017 size 100
[junit4] 2> 1568002 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741842_1018{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1568003 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741842_1018{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW], ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW]]} size 0
[junit4] 2> 1568018 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:37925 is added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW]]} size 74
[junit4] 2> 1568018 DEBUG (ScheduledTrigger-6910-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
[junit4] 2> 1568018 DEBUG (ScheduledTrigger-6910-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Tracking lost node: 127.0.0.1:38621_solr
[junit4] 2> 1568018 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:46851 is added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51:NORMAL:127.0.0.1:37925|RBW], ReplicaUC[[DISK]DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1:NORMAL:127.0.0.1:46851|RBW]]} size 74
[junit4] 2> 1568027 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsDirectory Closing hdfs directory hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data
[junit4] 2> 1568028 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsDirectory Closing hdfs directory hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/snapshot_metadata
[junit4] 2> 1568028 INFO (coreCloseExecutor-6929-thread-1) [n:127.0.0.1:38621_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsDirectory Closing hdfs directory hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/index
[junit4] 2> 1568031 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.c.Overseer Overseer (id=72089692485255172-127.0.0.1:38621_solr-n_0000000000) closing
[junit4] 2> 1568031 INFO (OverseerStateUpdate-72089692485255172-127.0.0.1:38621_solr-n_0000000000) [n:127.0.0.1:38621_solr ] o.a.s.c.Overseer Overseer Loop exiting : 127.0.0.1:38621_solr
[junit4] 2> 1568029 ERROR (OldIndexDirectoryCleanupThreadForCore-movereplicatest_coll3_shard1_replica_n1) [ ] o.a.s.c.HdfsDirectoryFactory Error checking for old index directories to clean-up.
[junit4] 2> java.io.IOException: Filesystem closed
[junit4] 2> at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:808) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2083) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2069) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:791) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517) ~[hadoop-common-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1557) ~[hadoop-common-2.7.4.jar:?]
[junit4] 2> at org.apache.solr.core.HdfsDirectoryFactory.cleanupOldIndexDirectories(HdfsDirectoryFactory.java:528) ~[java/:?]
[junit4] 2> at org.apache.solr.core.SolrCore.lambda$cleanupOldIndexDirectories$21(SolrCore.java:3099) ~[java/:?]
[junit4] 2> at java.lang.Thread.run(Thread.java:748) [?:1.8.0_172]
[junit4] 2> 1568031 WARN (OverseerAutoScalingTriggerThread-72089692485255172-127.0.0.1:38621_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
[junit4] 2> 1568037 INFO (zkCallback-3521-thread-1) [ ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:46313_solr
[junit4] 2> 1568037 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@ae680d{/solr,null,UNAVAILABLE}
[junit4] 2> 1568038 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.e.j.s.session node0 Stopped scavenging
[junit4] 2> 1568039 INFO (zkCallback-3521-thread-1) [n:127.0.0.1:46313_solr ] o.a.s.c.Overseer Overseer (id=72089692485255175-127.0.0.1:46313_solr-n_0000000001) starting
[junit4] 2> 1568048 INFO (qtp32863764-15885) [n:127.0.0.1:46313_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :movereplica with params replica=core_node2&action=MOVEREPLICA&collection=movereplicatest_coll3&targetNode=127.0.0.1:46313_solr&wt=javabin&version=2&inPlaceMove=true and sendToOCPQueue=true
[junit4] 2> 1568048 INFO (OverseerStateUpdate-72089692485255175-127.0.0.1:46313_solr-n_0000000001) [n:127.0.0.1:46313_solr ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:46313_solr
[junit4] 2> 1568053 INFO (OverseerThreadFactory-6934-thread-1-processing-n:127.0.0.1:46313_solr) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 r:core_node2 ] o.a.s.c.a.c.MoveReplicaCmd Replica will be moved to node 127.0.0.1:46313_solr: core_node2:{"dataDir":"hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/","base_url":"https://127.0.0.1:38621/solr","node_name":"127.0.0.1:38621_solr","type":"NRT","force_set_state":"false","ulogDir":"hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/tlog","core":"movereplicatest_coll3_shard1_replica_n1","shared_storage":"true","state":"down","leader":"true"}
[junit4] 2> 1568053 DEBUG (OverseerAutoScalingTriggerThread-72089692485255175-127.0.0.1:46313_solr-n_0000000001) [ ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:46313_solr]
[junit4] 2> 1568053 DEBUG (OverseerAutoScalingTriggerThread-72089692485255175-127.0.0.1:46313_solr-n_0000000001) [ ] o.a.s.c.a.NodeLostTrigger Adding lost node from marker path: 127.0.0.1:38621_solr
[junit4] 2> 1568053 INFO (OverseerThreadFactory-6934-thread-1-processing-n:127.0.0.1:46313_solr) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 r:core_node2 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:46313_solr for creating new replica of shard shard1 for collection movereplicatest_coll3
[junit4] 2> 1568055 DEBUG (ScheduledTrigger-6932-thread-2) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
[junit4] 2> 1568065 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr x:movereplicatest_coll3_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node2&dataDir=hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/&collection.configName=conf1&name=movereplicatest_coll3_shard1_replica_n1&action=CREATE&collection=movereplicatest_coll3&shard=shard1&wt=javabin&version=2&ulogDir=hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/&replicaType=NRT
[junit4] 2> 1569057 DEBUG (ScheduledTrigger-6932-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
[junit4] 2> 1569099 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1569113 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.IndexSchema [movereplicatest_coll3_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1569116 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1569116 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'movereplicatest_coll3_shard1_replica_n1' using configuration from collection movereplicatest_coll3, trusted=true
[junit4] 2> 1569117 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_46313.solr.core.movereplicatest_coll3.shard1.replica_n1' (registry 'solr.core.movereplicatest_coll3.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@ef4b43
[junit4] 2> 1569117 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost.localdomain:43569/data
[junit4] 2> 1569117 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 1569117 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrCore [[movereplicatest_coll3_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-001/node1/movereplicatest_coll3_shard1_replica_n1], dataDir=[hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/]
[junit4] 2> 1569120 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/snapshot_metadata
[junit4] 2> 1569133 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data
[junit4] 2> 1569223 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 1569223 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1569223 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
[junit4] 2> 1569235 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.FSHDFSUtils Recovering lease on dfs file hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/tlog/tlog.0000000000000000000
[junit4] 2> 1569334 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1569334 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1569337 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/index
[junit4] 2> 1569365 INFO (IPC Server handler 8 on 43569) [ ] BlockStateChange BLOCK* addToInvalidates: blk_1073741825_1001 127.0.0.1:37925 127.0.0.1:46851
[junit4] 2> 1569394 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@190117b[movereplicatest_coll3_shard1_replica_n1] main]
[junit4] 2> 1569396 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1569396 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1569397 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1569400 WARN (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1569400 INFO (searcherExecutor-6935-thread-1-processing-n:127.0.0.1:46313_solr x:movereplicatest_coll3_shard1_replica_n1 c:movereplicatest_coll3 s:shard1 r:core_node2) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrCore [movereplicatest_coll3_shard1_replica_n1] Registered new searcher Searcher@190117b[movereplicatest_coll3_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(7.6.0):C2)))}
[junit4] 2> 1569412 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1569412 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1569412 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:46313/solr/movereplicatest_coll3_shard1_replica_n1/
[junit4] 2> 1569413 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1569413 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:46313/solr/movereplicatest_coll3_shard1_replica_n1/ has no replicas
[junit4] 2> 1569413 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1569414 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:46313/solr/movereplicatest_coll3_shard1_replica_n1/ shard1
[junit4] 2> 1569426 WARN (Thread-4084) [ ] o.a.h.h.DFSClient DataStreamer Exception
[junit4] 2> java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[127.0.0.1:37925,DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51,DISK], DatanodeInfoWithStorage[127.0.0.1:46851,DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1,DISK]], original=[DatanodeInfoWithStorage[127.0.0.1:46851,DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1,DISK], DatanodeInfoWithStorage[127.0.0.1:37925,DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:1044) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1107) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1276) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:560) [hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> 1569428 ERROR (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.u.HdfsTransactionLog Could not close tlog output
[junit4] 2> java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[127.0.0.1:37925,DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51,DISK], DatanodeInfoWithStorage[127.0.0.1:46851,DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1,DISK]], original=[DatanodeInfoWithStorage[127.0.0.1:46851,DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1,DISK], DatanodeInfoWithStorage[127.0.0.1:37925,DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:1044) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1107) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1276) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:560) ~[hadoop-hdfs-2.7.4.jar:?]
[junit4] 2> 1569429 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1569469 INFO (qtp32863764-15882) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node2&dataDir=hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/&collection.configName=conf1&name=movereplicatest_coll3_shard1_replica_n1&action=CREATE&collection=movereplicatest_coll3&shard=shard1&wt=javabin&version=2&ulogDir=hdfs://localhost.localdomain:43569/data/movereplicatest_coll3/core_node2/data/&replicaType=NRT} status=0 QTime=1403
[junit4] 2> 1569473 INFO (qtp32863764-15885) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={replica=core_node2&action=MOVEREPLICA&collection=movereplicatest_coll3&targetNode=127.0.0.1:46313_solr&wt=javabin&version=2&inPlaceMove=true} status=0 QTime=1425
[junit4] 2> 1569532 INFO (zkCallback-3521-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/movereplicatest_coll3/state.json] for collection [movereplicatest_coll3] has occurred - updating... (live nodes size: [1])
[junit4] 2> 1569532 INFO (zkCallback-3521-thread-2) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/movereplicatest_coll3/state.json] for collection [movereplicatest_coll3] has occurred - updating... (live nodes size: [1])
[junit4] 2> 1570057 DEBUG (ScheduledTrigger-6932-thread-4) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
[junit4] 2> 1570057 INFO (OverseerCollectionConfigSetProcessor-72089692485255175-127.0.0.1:46313_solr-n_0000000001) [n:127.0.0.1:46313_solr ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 1570477 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.e.j.s.AbstractConnector Stopped ServerConnector@189d02a{SSL,[ssl, http/1.1]}{127.0.0.1:0}
[junit4] 2> 1570480 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.c.CoreContainer Shutting down CoreContainer instance=21311162
[junit4] 2> 1570481 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.node, tag=null
[junit4] 2> 1570481 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@b776c7: rootName = solr_46313, domain = solr.node, service url = null, agent id = null] for registry solr.node / com.codahale.metrics.MetricRegistry@173921b
[junit4] 2> 1570489 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.jvm, tag=null
[junit4] 2> 1570489 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@1b5c4d4: rootName = solr_46313, domain = solr.jvm, service url = null, agent id = null] for registry solr.jvm / com.codahale.metrics.MetricRegistry@b0f23c
[junit4] 2> 1570493 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.jetty, tag=null
[junit4] 2> 1570493 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@e5e360: rootName = solr_46313, domain = solr.jetty, service url = null, agent id = null] for registry solr.jetty / com.codahale.metrics.MetricRegistry@1d029e2
[junit4] 2> 1570495 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.c.ZkController Remove node as live in ZooKeeper:/live_nodes/127.0.0.1:46313_solr
[junit4] 2> 1570495 INFO (TEST-MoveReplicaHDFSFailoverTest.testOldReplicaIsDeleted-seed#[4270250C0D0814BA]) [ ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.cluster, tag=null
[junit4] 2> 1570495 INFO (zkCallback-3536-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (0)
[junit4] 2> 1570495 INFO (zkCallback-3521-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (0)
[junit4] 2> 1570496 INFO (zkCallback-3528-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (0)
[junit4] 2> 1570496 INFO (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.c.SolrCore [movereplicatest_coll3_shard1_replica_n1] CLOSING SolrCore org.apache.solr.core.SolrCore@8cb816
[junit4] 2> 1570496 INFO (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.movereplicatest_coll3.shard1.replica_n1, tag=8cb816
[junit4] 2> 1570497 INFO (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@15b2be4: rootName = solr_46313, domain = solr.core.movereplicatest_coll3.shard1.replica_n1, service url = null, agent id = null] for registry solr.core.movereplicatest_coll3.shard1.replica_n1 / com.codahale.metrics.MetricRegistry@19e5e3f
[junit4] 2> 1570505 WARN (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1570506 WARN (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1570506 WARN (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.s.h.HdfsLocalityReporter Could not retrieve locality information for hdfs://localhost.localdomain:44051/solr3 due to exception: java.net.ConnectException: Call From serv1.sd-datasolutions.de/88.99.242.108 to localhost.localdomain:44051 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
[junit4] 2> 1570513 INFO (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.movereplicatest_coll3.shard1.leader, tag=8cb816
[junit4] 2> 1570516 ERROR (coreCloseExecutor-6940-thread-1) [n:127.0.0.1:46313_solr c:movereplicatest_coll3 s:shard1 r:core_node2 x:movereplicatest_coll3_shard1_replica_n1] o.a.h.h.DFSClient Failed to close inode 16396
[junit4] 2> java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[127.0.0.1:37925,DS-e46c1a56-23aa-4c0a-b0a3-b0642afbff51,DISK], DatanodeInfoWithStorage[127.0.0.1:46851,DS-6ccdc4c0-6a85-4bc9-a6f3-bb7860705fc1,DISK]], original=[DatanodeInfoWithStorage[127.0.0.1:46
[...truncated too long message...]
alhost:0
[junit4] 2> 1643878 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:43569) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 1643878 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:43569) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-1489582878-88.99.242.108-1545552507144 (Datanode Uuid ec6c3b42-f20b-4412-9bd7-75c50b5af405) service to localhost.localdomain/127.0.0.1:43569
[junit4] 2> 1643882 WARN (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called
[junit4] 2> 1643889 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0
[junit4] 2> 1643990 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:43569) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted
[junit4] 2> 1643990 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:43569) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-1489582878-88.99.242.108-1545552507144 (Datanode Uuid 8ff5a187-be2d-435b-810c-df8b67206c6f) service to localhost.localdomain/127.0.0.1:43569
[junit4] 2> 1643999 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:0
[junit4] 2> 1644129 INFO (SUITE-MoveReplicaHDFSFailoverTest-seed#[4270250C0D0814BA]-worker) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:35221 35221
[junit4] 2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-7.6-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSFailoverTest_4270250C0D0814BA-001
[junit4] 2> Dec 23, 2018 8:09:48 AM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 65 leaked thread(s).
[junit4] 2> NOTE: test params are: codec=Asserting(Lucene70): {id=PostingsFormat(name=MockRandom)}, docValues:{_version_=DocValuesFormat(name=Asserting)}, maxPointsInLeafNode=1975, maxMBSortInHeap=5.413038810192915, sim=RandomSimilarity(queryNorm=true): {}, locale=fr-BE, timezone=America/Punta_Arenas
[junit4] 2> NOTE: Linux 4.15.0-42-generic i386/Oracle Corporation 1.8.0_172 (32-bit)/cpus=8,threads=3,free=167022784,total=428343296
[junit4] 2> NOTE: All tests run in this JVM: [TestQueryTypes, TestFuzzyAnalyzedSuggestions, TestFilteredDocIdSet, V2ApiIntegrationTest, HdfsUnloadDistributedZkTest, TestAuthorizationFramework, SolrCmdDistributorTest, TestJsonFacetRefinement, TestBinaryField, TestPerFieldSimilarity, TestCoreContainer, ResourceLoaderTest, CoreAdminCreateDiscoverTest, LeaderFailoverAfterPartitionTest, TestMinMaxOnMultiValuedField, TestSizeLimitedDistributedMap, TestSolrCloudWithHadoopAuthPlugin, TestStressLucene, TestHttpShardHandlerFactory, TestUpdate, ClassificationUpdateProcessorIntegrationTest, TestSolrDeletionPolicy2, BigEndianAscendingWordDeserializerTest, TestManagedSchemaThreadSafety, TestTrieFacet, TestManagedStopFilterFactory, OverriddenZkACLAndCredentialsProvidersTest, SpatialFilterTest, RandomizedTaggerTest, TestEmbeddedSolrServerAdminHandler, TestLeaderElectionZkExpiry, QueryEqualityTest, SolrJmxReporterCloudTest, SecurityConfHandlerTest, DistributedIntervalFacetingTest, TestExtendedDismaxParser, ZookeeperStatusHandlerTest, TestLegacyNumericRangeQueryBuilder, PeerSyncTest, TestHdfsCloudBackupRestore, BlockJoinFacetRandomTest, SuggestComponentTest, CollectionsAPISolrJTest, LeaderElectionContextKeyTest, TriggerSetPropertiesIntegrationTest, TestConfigSetImmutable, PhrasesIdentificationComponentTest, TestSha256AuthenticationProvider, TestConfigSetsAPIZkFailure, ParsingFieldUpdateProcessorsTest, SuggesterWFSTTest, ConvertedLegacyTest, AsyncCallRequestStatusResponseTest, TestWriterPerf, TestClusterProperties, TestSimExtremeIndexing, TestRandomRequestDistribution, TestCharFilters, TestFacetMethods, ByteBuffersDirectoryFactoryTest, ImplicitSnitchTest, SpellCheckCollatorTest, TestXIncludeConfig, TestConfigReload, TestChildDocTransformerHierarchy, ResponseHeaderTest, TestRecoveryHdfs, DeleteReplicaTest, StatelessScriptUpdateProcessorFactoryTest, BlockCacheTest, TestDynamicURP, HdfsDirectoryFactoryTest, TestDFISimilarityFactory, CoreAdminRequestStatusTest, TestStressInPlaceUpdates, TestDynamicFieldCollectionResource, TestMiniSolrCloudClusterSSL, TestLMDirichletSimilarityFactory, NodeAddedTriggerTest, HighlighterConfigTest, TestSmileRequest, DirectUpdateHandlerTest, TestSchemaManager, TaggingAttributeTest, TestSimDistribStateManager, TestTolerantUpdateProcessorCloud, CustomCollectionTest, ForceLeaderTest, ManagedSchemaRoundRobinCloudTest, TestCustomStream, JSONWriterTest, TestSolrCoreParser, HLLUtilTest, HttpPartitionTest, UtilsToolTest, TestSchemaVersionResource, TestFieldCacheWithThreads, DateMathParserTest, HdfsRecoverLeaseTest, DirectSolrSpellCheckerTest, SpellingQueryConverterTest, SolrCloudReportersTest, TestMultiWordSynonyms, TestCloudInspectUtil, TestSimNodeAddedTrigger, AlternateDirectoryTest, DebugComponentTest, MultiThreadedOCPTest, TestSimpleTrackingShardHandler, NumericFieldsTest, DirectSolrConnectionTest, SimpleCollectionCreateDeleteTest, TestMergePolicyConfig, StatsReloadRaceTest, DistributedFacetExistsSmallTest, MetricTriggerIntegrationTest, CdcrRequestHandlerTest, TestCloudSchemaless, TestCustomSort, DocValuesMultiTest, SolrXmlInZkTest, PingRequestHandlerTest, TestExactSharedStatsCache, TestJsonFacetsWithNestedObjects, UpdateParamsTest, TestNumericTerms64, TestExclusionRuleCollectionAccess, PreAnalyzedFieldManagedSchemaCloudTest, CleanupOldIndexTest, AnalysisErrorHandlingTest, TestCSVResponseWriter, SpatialHeatmapFacetsTest, ZkNodePropsTest, XsltUpdateRequestHandlerTest, AssignTest, AnalysisAfterCoreReloadTest, BasicFunctionalityTest, EchoParamsTest, MinimalSchemaTest, OutputWriterTest, SampleTest, SolrInfoBeanTest, SolrTestCaseJ4Test, TestDistributedGrouping, TestDocumentBuilder, TestGroupingSearch, TestHighlightDedupGrouping, TestJoin, TestRandomDVFaceting, TestRandomFaceting, TestSolrCoreProperties, TestTolerantSearch, TestTrie, TestWordDelimiterFilterFactory, TestEmbeddedSolrServerConstructors, TestEmbeddedSolrServerSchemaAPI, TestJettySolrRunner, ConnectionReuseTest, ActionThrottleTest, AddReplicaTest, BasicZkTest, ChaosMonkeySafeLeaderTest, CloudExitableDirectoryReaderTest, ClusterStateTest, ClusterStateUpdateTest, ConcurrentCreateRoutedAliasTest, ConfigSetsAPITest, ConnectionManagerTest, CreateCollectionCleanupTest, CreateRoutedAliasTest, DeleteInactiveReplicaTest, DeleteLastCustomShardedReplicaTest, DeleteNodeTest, DistribCursorPagingTest, DistributedQueueTest, DistributedVersionInfoTest, DocValuesNotIndexedTest, LeaderVoteWaitTimeoutTest, LegacyCloudClusterPropTest, MetricsHistoryIntegrationTest, MigrateRouteKeyTest, MissingSegmentRecoveryTest, MoveReplicaHDFSFailoverTest]
[junit4] Completed [446/836 (1!)] on J2 in 92.23s, 3 tests, 1 error <<< FAILURES!
[...truncated 46632 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-7.6-Linux/139/consoleText
[repro] Revision: e1d5761f7b976aa4ab83969f9a699597c0855b3e
[repro] Ant options: "-Dargs=-client -XX:+UseParallelGC"
[repro] Repro line: ant test -Dtestcase=MoveReplicaHDFSFailoverTest -Dtests.method=testOldReplicaIsDeletedInRaceCondition -Dtests.seed=4270250C0D0814BA -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=fr-BE -Dtests.timezone=America/Punta_Arenas -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[repro] ant clean
[...truncated 6 lines...]
[repro] Test suites by module:
[repro] solr/core
[repro] MoveReplicaHDFSFailoverTest
[repro] ant compile-test
[...truncated 3580 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.MoveReplicaHDFSFailoverTest" -Dtests.showOutput=onerror "-Dargs=-client -XX:+UseParallelGC" -Dtests.seed=4270250C0D0814BA -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=fr-BE -Dtests.timezone=America/Punta_Arenas -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[...truncated 81 lines...]
[repro] Failures:
[repro] 0/5 failed: org.apache.solr.cloud.MoveReplicaHDFSFailoverTest
[repro] Exiting with code 0
[...truncated 40 lines...]