You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@lucene.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/07/12 16:12:24 UTC

[JENKINS] Lucene » Lucene-Solr-Tests-8.9 - Build # 194 - Unstable!

Build: https://ci-builds.apache.org/job/Lucene/job/Lucene-Solr-Tests-8.9/194/

1 tests failed.
FAILED:  org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore.testRestoreFailure

Error Message:
Failed collection is still in the clusterstate: null Expected: not a collection containing "hdfsbackuprestore_testfailure_restored"      but: was <[hdfsbackuprestore_testfailure_restored, hdfsbackuprestore_testfailure]>

Stack Trace:
java.lang.AssertionError: Failed collection is still in the clusterstate: null
Expected: not a collection containing "hdfsbackuprestore_testfailure_restored"
     but: was <[hdfsbackuprestore_testfailure_restored, hdfsbackuprestore_testfailure]>
	at __randomizedtesting.SeedInfo.seed([1DF421E3E00928AB:3488BFC6C8502B86]:0)
	at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
	at org.junit.Assert.assertThat(Assert.java:964)
	at org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.testRestoreFailure(AbstractCloudBackupRestoreTestCase.java:223)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.junit.rules.RunRules.evaluate(RunRules.java:20)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.junit.rules.RunRules.evaluate(RunRules.java:20)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.junit.rules.RunRules.evaluate(RunRules.java:20)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at org.junit.rules.RunRules.evaluate(RunRules.java:20)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 14372 lines...]
   [junit4] Suite: org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore
   [junit4]   2> 828502 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.SolrTestCase Setting 'solr.default.confdir' system property to test-framework derived value of '/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/server/solr/configsets/_default/conf'
   [junit4]   2> 828502 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.SolrTestCaseJ4 Created dataDir: /home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/data-dir-52-001
   [junit4]   2> 828503 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=9 numCloses=9
   [junit4]   2> 828503 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=false
   [junit4]   2> 828504 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason=, value=NaN, ssl=NaN, clientAuth=NaN)
   [junit4]   2> 828505 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 828505 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.SolrCloudTestCase Using per-replica state
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 828549 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 828551 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.Server jetty-9.4.41.v20210516; built: 2021-05-16T23:56:28.993Z; git: 98607f93c7833e7dc59489b13f3cb0a114fb9f4c; jvm 1.8.0_291-b10
   [junit4]   2> 828561 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 828561 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 828561 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 828562 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@4f7f006c{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 828670 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.h.ContextHandler Started o.e.j.w.WebAppContext@3a20d08e{hdfs,/,file:///home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/jetty-localhost-46037-hadoop-hdfs-3_2_0-tests_jar-_-any-1667938912301253838/webapp/,AVAILABLE}{jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/hdfs}
   [junit4]   2> 828671 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.AbstractConnector Started ServerConnector@417d1c0a{HTTP/1.1, (http/1.1)}{localhost:46037}
   [junit4]   2> 828671 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.Server Started @828699ms
   [junit4]   2> 828733 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 828734 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.Server jetty-9.4.41.v20210516; built: 2021-05-16T23:56:28.993Z; git: 98607f93c7833e7dc59489b13f3cb0a114fb9f4c; jvm 1.8.0_291-b10
   [junit4]   2> 828737 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 828737 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 828738 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 828738 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@72b43cfe{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 828841 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.h.ContextHandler Started o.e.j.w.WebAppContext@1f0b4c51{datanode,/,file:///home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/jetty-localhost-42365-hadoop-hdfs-3_2_0-tests_jar-_-any-2001854617749171314/webapp/,AVAILABLE}{jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/datanode}
   [junit4]   2> 828841 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.AbstractConnector Started ServerConnector@3887a30b{HTTP/1.1, (http/1.1)}{localhost:42365}
   [junit4]   2> 828841 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.e.j.s.Server Started @828870ms
   [junit4]   2> 828953 INFO  (Block report processor) [     ] BlockStateChange BLOCK* processReport 0xb75cbdb2d7525872: Processing first storage report for DS-ea496fbb-3ae1-4c8f-af82-13dfa400b81e from datanode e2bf43e2-dfad-4f8a-804b-34ce2b4d26c9
   [junit4]   2> 828953 INFO  (Block report processor) [     ] BlockStateChange BLOCK* processReport 0xb75cbdb2d7525872: from storage DS-ea496fbb-3ae1-4c8f-af82-13dfa400b81e node DatanodeRegistration(127.0.0.1:36663, datanodeUuid=e2bf43e2-dfad-4f8a-804b-34ce2b4d26c9, infoPort=44727, infoSecurePort=0, ipcPort=40363, storageInfo=lv=-57;cid=testClusterID;nsid=352103623;c=1626103361353), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 828953 INFO  (Block report processor) [     ] BlockStateChange BLOCK* processReport 0xb75cbdb2d7525872: Processing first storage report for DS-6f211a58-df8b-422b-b118-292a91cc65ea from datanode e2bf43e2-dfad-4f8a-804b-34ce2b4d26c9
   [junit4]   2> 828953 INFO  (Block report processor) [     ] BlockStateChange BLOCK* processReport 0xb75cbdb2d7525872: from storage DS-6f211a58-df8b-422b-b118-292a91cc65ea node DatanodeRegistration(127.0.0.1:36663, datanodeUuid=e2bf43e2-dfad-4f8a-804b-34ce2b4d26c9, infoPort=44727, infoSecurePort=0, ipcPort=40363, storageInfo=lv=-57;cid=testClusterID;nsid=352103623;c=1626103361353), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 828990 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.a.c.TestHdfsCloudBackupRestore The NameNode is in SafeMode - Solr will wait 5 seconds and try again.
   [junit4]   2> 833991 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.a.c.TestHdfsCloudBackupRestore The NameNode is in SafeMode - Solr will wait 5 seconds and try again.
   [junit4]   2> 838994 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in /home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002
   [junit4]   2> 838994 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 838994 INFO  (ZkTestServer Run Thread) [     ] o.a.s.c.ZkTestServer client port: 0.0.0.0/0.0.0.0:0
   [junit4]   2> 838994 INFO  (ZkTestServer Run Thread) [     ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 838995 WARN  (ZkTestServer Run Thread) [     ] o.a.z.s.ServerCnxnFactory maxCnxns is not configured, using default value 0.
   [junit4]   2> 839094 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer start zk server on port: 34839
   [junit4]   2> 839095 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer waitForServerUp: 127.0.0.1:34839
   [junit4]   2> 839095 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer parse host and port list: 127.0.0.1:34839
   [junit4]   2> 839095 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer connecting to 127.0.0.1 34839
   [junit4]   2> 839096 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 839101 INFO  (zkConnectionManagerCallback-5758-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 839101 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 839104 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 839113 INFO  (zkConnectionManagerCallback-5760-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 839113 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 839114 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 839116 INFO  (zkConnectionManagerCallback-5762-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 839116 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 839126 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker-SendThread(127.0.0.1:34839)) [     ] o.a.z.ClientCnxn An exception was thrown while closing send thread for session 0x106523d7d6e0002.
   [junit4]   2>           => EndOfStreamException: Unable to read additional data from server sessionid 0x106523d7d6e0002, likely server has closed socket
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77)
   [junit4]   2> org.apache.zookeeper.ClientCnxn$EndOfStreamException: Unable to read additional data from server sessionid 0x106523d7d6e0002, likely server has closed socket
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77) ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1275) [zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 839234 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.s.e.JettySolrRunner Start Jetty (configured port=0, binding port=0)
   [junit4]   2> 839234 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 2 ...
   [junit4]   2> 839234 INFO  (jetty-launcher-5763-thread-1) [     ] o.e.j.s.Server jetty-9.4.41.v20210516; built: 2021-05-16T23:56:28.993Z; git: 98607f93c7833e7dc59489b13f3cb0a114fb9f4c; jvm 1.8.0_291-b10
   [junit4]   2> 839236 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.s.e.JettySolrRunner Start Jetty (configured port=0, binding port=0)
   [junit4]   2> 839236 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 2 ...
   [junit4]   2> 839236 INFO  (jetty-launcher-5763-thread-2) [     ] o.e.j.s.Server jetty-9.4.41.v20210516; built: 2021-05-16T23:56:28.993Z; git: 98607f93c7833e7dc59489b13f3cb0a114fb9f4c; jvm 1.8.0_291-b10
   [junit4]   2> 839256 INFO  (jetty-launcher-5763-thread-1) [     ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 839256 INFO  (jetty-launcher-5763-thread-1) [     ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 839257 INFO  (jetty-launcher-5763-thread-2) [     ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 839263 INFO  (jetty-launcher-5763-thread-2) [     ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 839263 INFO  (jetty-launcher-5763-thread-2) [     ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 839263 INFO  (jetty-launcher-5763-thread-1) [     ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 839267 INFO  (jetty-launcher-5763-thread-2) [     ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@357fa55f{/solr,null,AVAILABLE}
   [junit4]   2> 839270 INFO  (jetty-launcher-5763-thread-1) [     ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3e102a18{/solr,null,AVAILABLE}
   [junit4]   2> 839271 INFO  (jetty-launcher-5763-thread-2) [     ] o.e.j.s.AbstractConnector Started ServerConnector@3af04e43{SSL, (ssl, http/1.1)}{127.0.0.1:32795}
   [junit4]   2> 839271 INFO  (jetty-launcher-5763-thread-2) [     ] o.e.j.s.Server Started @839300ms
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, zkHost=127.0.0.1:34839/solr, hostPort=32795}
   [junit4]   2> 839272 ERROR (jetty-launcher-5763-thread-2) [     ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-1) [     ] o.e.j.s.AbstractConnector Started ServerConnector@73f8341a{SSL, (ssl, http/1.1)}{127.0.0.1:33109}
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.9.1
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: /home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2021-07-12T15:22:52.109Z
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-1) [     ] o.e.j.s.Server Started @839300ms
   [junit4]   2> 839272 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, zkHost=127.0.0.1:34839/solr, hostPort=33109}
   [junit4]   2> 839273 ERROR (jetty-launcher-5763-thread-1) [     ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 839273 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 839273 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.9.1
   [junit4]   2> 839273 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 839273 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: /home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr
   [junit4]   2> 839273 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2021-07-12T15:22:52.110Z
   [junit4]   2> 839280 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 839285 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 839289 INFO  (zkConnectionManagerCallback-5765-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 839289 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 839290 INFO  (zkConnectionManagerCallback-5767-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 839291 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 839291 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 839292 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 839296 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 839296 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 839741 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized: WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
   [junit4]   2> 839741 WARN  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 839744 WARN  (jetty-launcher-5763-thread-2) [     ] o.e.j.u.s.S.config Trusting all certificates configured for Client@5853a3d8[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 839744 WARN  (jetty-launcher-5763-thread-2) [     ] o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for Client@5853a3d8[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 839750 WARN  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 839750 WARN  (jetty-launcher-5763-thread-2) [     ] o.e.j.u.s.S.config Trusting all certificates configured for Client@50598fb3[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 839751 WARN  (jetty-launcher-5763-thread-2) [     ] o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for Client@50598fb3[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 839752 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:34839/solr
   [junit4]   2> 839753 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 839757 INFO  (zkConnectionManagerCallback-5781-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 839757 INFO  (jetty-launcher-5763-thread-2) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 839864 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 839865 INFO  (zkConnectionManagerCallback-5783-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 839865 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 839927 WARN  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.ZkController Contents of zookeeper /security.json are world-readable; consider setting up ACLs as described in https://solr.apache.org/guide/zookeeper-access-control.html
   [junit4]   2> 839930 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:32795_solr
   [junit4]   2> 839930 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.Overseer Overseer (id=73836867949035526-127.0.0.1:32795_solr-n_0000000000) starting
   [junit4]   2> 839934 INFO  (OverseerStateUpdate-73836867949035526-127.0.0.1:32795_solr-n_0000000000) [n:127.0.0.1:32795_solr     ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:32795_solr
   [junit4]   2> 839937 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:32795_solr
   [junit4]   2> 839940 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.p.PackageLoader /packages.json updated to version -1
   [junit4]   2> 839940 WARN  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.CoreContainer Not all security plugins configured!  authentication=disabled authorization=disabled.  Solr is only as secure as you make it. Consider configuring authentication/authorization before exposing Solr to users internal or external.  See https://s.apache.org/solrsecurity for more info
   [junit4]   2> 839940 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup repository with configuration params {type = repository,name = hdfs,class = org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = {name=hdfs, class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = {location=/backup,solr.hdfs.home=hdfs://localhost:37773/solr,solr.hdfs.confdir=}}
   [junit4]   2> 839940 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup repository with configuration params {type = repository,name = poisioned,class = org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes = {default=true, name=poisioned, class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 839940 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Default configuration for backup repository is with configuration params {type = repository,name = poisioned,class = org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes = {default=true, name=poisioned, class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 839946 INFO  (zkCallback-5782-thread-1) [     ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 839969 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 839991 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 839999 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 839999 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840000 INFO  (jetty-launcher-5763-thread-2) [n:127.0.0.1:32795_solr     ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node2
   [junit4]   2> 840137 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized: WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
   [junit4]   2> 840137 WARN  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 840139 WARN  (jetty-launcher-5763-thread-1) [     ] o.e.j.u.s.S.config Trusting all certificates configured for Client@4294a2a4[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 840139 WARN  (jetty-launcher-5763-thread-1) [     ] o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for Client@4294a2a4[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 840141 WARN  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 840145 WARN  (jetty-launcher-5763-thread-1) [     ] o.e.j.u.s.S.config Trusting all certificates configured for Client@72d6b559[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 840145 WARN  (jetty-launcher-5763-thread-1) [     ] o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for Client@72d6b559[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 840147 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:34839/solr
   [junit4]   2> 840147 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 840149 INFO  (zkConnectionManagerCallback-5803-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 840149 INFO  (jetty-launcher-5763-thread-1) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 840251 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 840255 INFO  (zkConnectionManagerCallback-5805-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 840255 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 840259 WARN  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.ZkController Contents of zookeeper /security.json are world-readable; consider setting up ACLs as described in https://solr.apache.org/guide/zookeeper-access-control.html
   [junit4]   2> 840260 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 840262 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.ZkController Publish node=127.0.0.1:33109_solr as DOWN
   [junit4]   2> 840263 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 840263 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:33109_solr
   [junit4]   2> 840264 INFO  (zkCallback-5782-thread-1) [     ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 840265 INFO  (zkCallback-5804-thread-1) [     ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 840266 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.p.PackageLoader /packages.json updated to version -1
   [junit4]   2> 840266 WARN  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.CoreContainer Not all security plugins configured!  authentication=disabled authorization=disabled.  Solr is only as secure as you make it. Consider configuring authentication/authorization before exposing Solr to users internal or external.  See https://s.apache.org/solrsecurity for more info
   [junit4]   2> 840266 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup repository with configuration params {type = repository,name = hdfs,class = org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = {name=hdfs, class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = {location=/backup,solr.hdfs.home=hdfs://localhost:37773/solr,solr.hdfs.confdir=}}
   [junit4]   2> 840266 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup repository with configuration params {type = repository,name = poisioned,class = org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes = {default=true, name=poisioned, class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 840266 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Default configuration for backup repository is with configuration params {type = repository,name = poisioned,class = org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes = {default=true, name=poisioned, class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 840277 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 840296 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840306 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840306 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840309 INFO  (jetty-launcher-5763-thread-1) [n:127.0.0.1:33109_solr     ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node1
   [junit4]   2> 840330 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.MiniSolrCloudCluster waitForAllNodes: numServers=2
   [junit4]   2> 840331 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 840332 INFO  (zkConnectionManagerCallback-5818-thread-1) [     ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 840332 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 840334 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 840335 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:34839/solr ready
   [junit4]   2> 840382 INFO  (TEST-TestHdfsCloudBackupRestore.testRestoreFailure-seed#[1DF421E3E00928AB]) [     ] o.a.s.SolrTestCaseJ4 ###Starting testRestoreFailure
   [junit4]   2> 840414 INFO  (OverseerThreadFactory-5790-thread-1-processing-n:127.0.0.1:32795_solr) [n:127.0.0.1:32795_solr     ] o.a.s.c.a.c.CreateCollectionCmd Create collection hdfsbackuprestore_testfailure
   [junit4]   2> 840518 WARN  (OverseerThreadFactory-5790-thread-1-processing-n:127.0.0.1:32795_solr) [n:127.0.0.1:32795_solr     ] o.a.s.c.a.c.CreateCollectionCmd Specified number of replicas of 3 on collection hdfsbackuprestore_testfailure is higher than the number of Solr instances currently live or live and part of your createNodeSet(2). It's unusual to run two replica of the same slice on the same Solr-instance.
   [junit4]   2> 840522 INFO  (OverseerStateUpdate-73836867949035526-127.0.0.1:32795_solr-n_0000000000) [n:127.0.0.1:32795_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "core":"hdfsbackuprestore_testfailure_shard1_replica_n1",
   [junit4]   2>   "node_name":"127.0.0.1:33109_solr",
   [junit4]   2>   "base_url":"https://127.0.0.1:33109/solr",
   [junit4]   2>   "collection":"hdfsbackuprestore_testfailure",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 840524 INFO  (OverseerStateUpdate-73836867949035526-127.0.0.1:32795_solr-n_0000000000) [n:127.0.0.1:32795_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "core":"hdfsbackuprestore_testfailure_shard1_replica_n2",
   [junit4]   2>   "node_name":"127.0.0.1:32795_solr",
   [junit4]   2>   "base_url":"https://127.0.0.1:32795/solr",
   [junit4]   2>   "collection":"hdfsbackuprestore_testfailure",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 840527 INFO  (OverseerStateUpdate-73836867949035526-127.0.0.1:32795_solr-n_0000000000) [n:127.0.0.1:32795_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "core":"hdfsbackuprestore_testfailure_shard1_replica_t4",
   [junit4]   2>   "node_name":"127.0.0.1:33109_solr",
   [junit4]   2>   "base_url":"https://127.0.0.1:33109/solr",
   [junit4]   2>   "collection":"hdfsbackuprestore_testfailure",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "type":"TLOG",
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 840529 INFO  (OverseerStateUpdate-73836867949035526-127.0.0.1:32795_solr-n_0000000000) [n:127.0.0.1:32795_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "core":"hdfsbackuprestore_testfailure_shard2_replica_n6",
   [junit4]   2>   "node_name":"127.0.0.1:32795_solr",
   [junit4]   2>   "base_url":"https://127.0.0.1:32795/solr",
   [junit4]   2>   "collection":"hdfsbackuprestore_testfailure",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 840531 INFO  (OverseerStateUpdate-73836867949035526-127.0.0.1:32795_solr-n_0000000000) [n:127.0.0.1:32795_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "core":"hdfsbackuprestore_testfailure_shard2_replica_n8",
   [junit4]   2>   "node_name":"127.0.0.1:33109_solr",
   [junit4]   2>   "base_url":"https://127.0.0.1:33109/solr",
   [junit4]   2>   "collection":"hdfsbackuprestore_testfailure",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 840533 INFO  (OverseerStateUpdate-73836867949035526-127.0.0.1:32795_solr-n_0000000000) [n:127.0.0.1:32795_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "core":"hdfsbackuprestore_testfailure_shard2_replica_t10",
   [junit4]   2>   "node_name":"127.0.0.1:32795_solr",
   [junit4]   2>   "base_url":"https://127.0.0.1:32795/solr",
   [junit4]   2>   "collection":"hdfsbackuprestore_testfailure",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "type":"TLOG",
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 840755 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr    x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard1_replica_n1&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 840756 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr    x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard1_replica_t4&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard1&wt=javabin&version=2&replicaType=TLOG
   [junit4]   2> 840757 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr    x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node11&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard2_replica_n8&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 840779 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.9.1
   [junit4]   2> 840783 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.9.1
   [junit4]   2> 840784 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 840785 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 840785 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.CoreContainer Creating SolrCore 'hdfsbackuprestore_testfailure_shard1_replica_t4' using configuration from configset conf1, trusted=true
   [junit4]   2> 840785 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.hdfsbackuprestore_testfailure.shard1.replica_t4' (registry 'solr.core.hdfsbackuprestore_testfailure.shard1.replica_t4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840785 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.SolrCore [[hdfsbackuprestore_testfailure_shard1_replica_t4] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node1/hdfsbackuprestore_testfailure_shard1_replica_t4], dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node1/hdfsbackuprestore_testfailure_shard1_replica_t4/data/]
   [junit4]   2> 840786 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr    x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node9&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard2_replica_n6&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 840787 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr    x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 840787 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 840788 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 840788 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.CoreContainer Creating SolrCore 'hdfsbackuprestore_testfailure_shard2_replica_n8' using configuration from configset conf1, trusted=true
   [junit4]   2> 840788 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.hdfsbackuprestore_testfailure.shard2.replica_n8' (registry 'solr.core.hdfsbackuprestore_testfailure.shard2.replica_n8') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840788 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.SolrCore [[hdfsbackuprestore_testfailure_shard2_replica_n8] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node1/hdfsbackuprestore_testfailure_shard2_replica_n8], dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node1/hdfsbackuprestore_testfailure_shard2_replica_n8/data/]
   [junit4]   2> 840810 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.9.1
   [junit4]   2> 840810 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr    x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node12&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard2_replica_t10&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard2&wt=javabin&version=2&replicaType=TLOG
   [junit4]   2> 840810 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr    x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard1_replica_n2&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 840811 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 840812 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 840813 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.c.CoreContainer Creating SolrCore 'hdfsbackuprestore_testfailure_shard1_replica_n1' using configuration from configset conf1, trusted=true
   [junit4]   2> 840813 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.hdfsbackuprestore_testfailure.shard1.replica_n1' (registry 'solr.core.hdfsbackuprestore_testfailure.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840813 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.c.SolrCore [[hdfsbackuprestore_testfailure_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node1/hdfsbackuprestore_testfailure_shard1_replica_n1], dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node1/hdfsbackuprestore_testfailure_shard1_replica_n1/data/]
   [junit4]   2> 840841 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.9.1
   [junit4]   2> 840849 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 840849 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 840849 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.CoreContainer Creating SolrCore 'hdfsbackuprestore_testfailure_shard2_replica_n6' using configuration from configset conf1, trusted=true
   [junit4]   2> 840850 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.hdfsbackuprestore_testfailure.shard2.replica_n6' (registry 'solr.core.hdfsbackuprestore_testfailure.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840850 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.SolrCore [[hdfsbackuprestore_testfailure_shard2_replica_n6] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node2/hdfsbackuprestore_testfailure_shard2_replica_n6], dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node2/hdfsbackuprestore_testfailure_shard2_replica_n6/data/]
   [junit4]   2> 840853 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.9.1
   [junit4]   2> 840853 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.9.1
   [junit4]   2> 840855 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 840855 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 840856 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 840856 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 840856 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.CoreContainer Creating SolrCore 'hdfsbackuprestore_testfailure_shard2_replica_t10' using configuration from configset conf1, trusted=true
   [junit4]   2> 840856 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.c.CoreContainer Creating SolrCore 'hdfsbackuprestore_testfailure_shard1_replica_n2' using configuration from configset conf1, trusted=true
   [junit4]   2> 840856 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.hdfsbackuprestore_testfailure.shard1.replica_n2' (registry 'solr.core.hdfsbackuprestore_testfailure.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840856 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.hdfsbackuprestore_testfailure.shard2.replica_t10' (registry 'solr.core.hdfsbackuprestore_testfailure.shard2.replica_t10') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bb7fff4
   [junit4]   2> 840856 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.c.SolrCore [[hdfsbackuprestore_testfailure_shard1_replica_n2] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node2/hdfsbackuprestore_testfailure_shard1_replica_n2], dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node2/hdfsbackuprestore_testfailure_shard1_replica_n2/data/]
   [junit4]   2> 840856 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.SolrCore [[hdfsbackuprestore_testfailure_shard2_replica_t10] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node2/hdfsbackuprestore_testfailure_shard2_replica_t10], dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001/tempDir-002/node2/hdfsbackuprestore_testfailure_shard2_replica_t10/data/]
   [junit4]   2> 840944 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 840944 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 840945 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 840945 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 840951 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 840951 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 840952 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 840952 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 840979 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 840979 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 840980 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 840981 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 840985 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 840992 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 840992 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 840992 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 840992 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 840992 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1705092971316117504
   [junit4]   2> 840994 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 840995 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 840995 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 840995 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1705092971319263232
   [junit4]   2> 840995 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 840995 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1705092971319263232
   [junit4]   2> 840995 INFO  (searcherExecutor-5824-thread-1-processing-n:127.0.0.1:33109_solr x:hdfsbackuprestore_testfailure_shard1_replica_n1 c:hdfsbackuprestore_testfailure s:shard1 r:core_node3) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.c.SolrCore [hdfsbackuprestore_testfailure_shard1_replica_n1]  Registered new searcher autowarm time: 0 ms
   [junit4]   2> 841002 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 841002 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 841002 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 841003 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 841003 INFO  (searcherExecutor-5820-thread-1-processing-n:127.0.0.1:33109_solr x:hdfsbackuprestore_testfailure_shard1_replica_t4 c:hdfsbackuprestore_testfailure s:shard1 r:core_node7) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.SolrCore [hdfsbackuprestore_testfailure_shard1_replica_t4]  Registered new searcher autowarm time: 0 ms
   [junit4]   2> 841004 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 841004 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 841004 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 841004 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 841009 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 841009 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 841009 INFO  (searcherExecutor-5822-thread-1-processing-n:127.0.0.1:33109_solr x:hdfsbackuprestore_testfailure_shard2_replica_n8 c:hdfsbackuprestore_testfailure s:shard2 r:core_node11) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.SolrCore [hdfsbackuprestore_testfailure_shard2_replica_n8]  Registered new searcher autowarm time: 0 ms
   [junit4]   2> 841010 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 841010 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 841013 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 841013 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 841014 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 841014 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1705092971339186176
   [junit4]   2> 841018 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 841020 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ZkShardTerms Successful update of terms at /collections/hdfsbackuprestore_testfailure/terms/shard1 to Terms{values={core_node7=0}, version=0}
   [junit4]   2> 841020 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ShardLeaderElectionContextBase make sure parent is created /collections/hdfsbackuprestore_testfailure/leaders/shard1
   [junit4]   2> 841020 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 841020 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 841021 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 841021 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1705092971346526208
   [junit4]   2> 841021 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 841022 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 841022 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1705092971347574784
   [junit4]   2> 841024 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ZkShardTerms Successful update of terms at /collections/hdfsbackuprestore_testfailure/terms/shard2 to Terms{values={core_node11=0}, version=0}
   [junit4]   2> 841024 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ShardLeaderElectionContextBase make sure parent is created /collections/hdfsbackuprestore_testfailure/leaders/shard2
   [junit4]   2> 841025 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.c.ZkShardTerms Successful update of terms at /collections/hdfsbackuprestore_testfailure/terms/shard1 to Terms{values={core_node7=0, core_node3=0}, version=1}
   [junit4]   2> 841025 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.c.ShardLeaderElectionContextBase make sure parent is created /collections/hdfsbackuprestore_testfailure/leaders/shard1
   [junit4]   2> 841027 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.ZkShardTerms Successful update of terms at /collections/hdfsbackuprestore_testfailure/terms/shard2 to Terms{values={core_node11=0, core_node9=0}, version=1}
   [junit4]   2> 841027 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.ShardLeaderElectionContextBase make sure parent is created /collections/hdfsbackuprestore_testfailure/leaders/shard2
   [junit4]   2> 841029 INFO  (searcherExecutor-5826-thread-1-processing-n:127.0.0.1:32795_solr x:hdfsbackuprestore_testfailure_shard2_replica_n6 c:hdfsbackuprestore_testfailure s:shard2 r:core_node9) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.SolrCore [hdfsbackuprestore_testfailure_shard2_replica_n6]  Registered new searcher autowarm time: 0 ms
   [junit4]   2> 841042 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard2: total=3 found=2 timeoute in=9989ms
   [junit4]   2> 841045 INFO  (searcherExecutor-5829-thread-1-processing-n:127.0.0.1:32795_solr x:hdfsbackuprestore_testfailure_shard1_replica_n2 c:hdfsbackuprestore_testfailure s:shard1 r:core_node5) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.c.SolrCore [hdfsbackuprestore_testfailure_shard1_replica_n2]  Registered new searcher autowarm time: 0 ms
   [junit4]   2> 841045 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.ZkShardTerms Successful update of terms at /collections/hdfsbackuprestore_testfailure/terms/shard2 to Terms{values={core_node12=0, core_node11=0, core_node9=0}, version=2}
   [junit4]   2> 841045 INFO  (searcherExecutor-5830-thread-1-processing-n:127.0.0.1:32795_solr x:hdfsbackuprestore_testfailure_shard2_replica_t10 c:hdfsbackuprestore_testfailure s:shard2 r:core_node12) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.SolrCore [hdfsbackuprestore_testfailure_shard2_replica_t10]  Registered new searcher autowarm time: 0 ms
   [junit4]   2> 841045 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.ShardLeaderElectionContextBase make sure parent is created /collections/hdfsbackuprestore_testfailure/leaders/shard2
   [junit4]   2> 841046 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard1: total=3 found=2 timeoute in=9998ms
   [junit4]   2> 841049 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.c.ZkShardTerms Successful update of terms at /collections/hdfsbackuprestore_testfailure/terms/shard1 to Terms{values={core_node7=0, core_node3=0, core_node5=0}, version=2}
   [junit4]   2> 841050 INFO  (qtp805840179-12299) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.c.ShardLeaderElectionContextBase make sure parent is created /collections/hdfsbackuprestore_testfailure/leaders/shard1
   [junit4]   2> 841543 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 841543 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 841543 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:33109/solr/hdfsbackuprestore_testfailure_shard2_replica_n8/
   [junit4]   2> 841544 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.u.PeerSync PeerSync: core=hdfsbackuprestore_testfailure_shard2_replica_n8 url=https://127.0.0.1:33109/solr  START replicas=[https://127.0.0.1:32795/solr/hdfsbackuprestore_testfailure_shard2_replica_n6/, https://127.0.0.1:32795/solr/hdfsbackuprestore_testfailure_shard2_replica_t10/] nUpdates=100
   [junit4]   2> 841544 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.u.PeerSync PeerSync: core=hdfsbackuprestore_testfailure_shard2_replica_n8 url=https://127.0.0.1:33109/solr  DONE. We have no versions. sync failed.
   [junit4]   2> 841554 INFO  (qtp805840179-12294) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.S.Request [hdfsbackuprestore_testfailure_shard2_replica_t10]  webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=1
   [junit4]   2> 841555 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 841555 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 841555 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:33109/solr/hdfsbackuprestore_testfailure_shard1_replica_t4/
   [junit4]   2> 841556 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.u.PeerSync PeerSync: core=hdfsbackuprestore_testfailure_shard1_replica_t4 url=https://127.0.0.1:33109/solr  START replicas=[https://127.0.0.1:33109/solr/hdfsbackuprestore_testfailure_shard1_replica_n1/, https://127.0.0.1:32795/solr/hdfsbackuprestore_testfailure_shard1_replica_n2/] nUpdates=100
   [junit4]   2> 841556 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.u.PeerSync PeerSync: core=hdfsbackuprestore_testfailure_shard1_replica_t4 url=https://127.0.0.1:33109/solr  DONE. We have no versions. sync failed.
   [junit4]   2> 841557 INFO  (qtp805840179-12294) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node9 x:hdfsbackuprestore_testfailure_shard2_replica_n6 ] o.a.s.c.S.Request [hdfsbackuprestore_testfailure_shard2_replica_n6]  webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 841561 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 841561 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 841561 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ShardLeaderElectionContextBase Creating leader registration node /collections/hdfsbackuprestore_testfailure/leaders/shard2/leader after winning as /collections/hdfsbackuprestore_testfailure/leader_elect/shard2/election/73836867949035528-core_node11-n_0000000000
   [junit4]   2> 841563 INFO  (qtp805840179-12295) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node5 x:hdfsbackuprestore_testfailure_shard1_replica_n2 ] o.a.s.c.S.Request [hdfsbackuprestore_testfailure_shard1_replica_n2]  webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 841565 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:33109/solr/hdfsbackuprestore_testfailure_shard2_replica_n8/ shard2
   [junit4]   2> 841569 INFO  (qtp2030067562-12289) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node3 x:hdfsbackuprestore_testfailure_shard1_replica_n1 ] o.a.s.c.S.Request [hdfsbackuprestore_testfailure_shard1_replica_n1]  webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 841570 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 841570 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 841570 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ZkController hdfsbackuprestore_testfailure_shard1_replica_t4 stopping background replication from leader
   [junit4]   2> 841570 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ShardLeaderElectionContextBase Creating leader registration node /collections/hdfsbackuprestore_testfailure/leaders/shard1/leader after winning as /collections/hdfsbackuprestore_testfailure/leader_elect/shard1/election/73836867949035528-core_node7-n_0000000000
   [junit4]   2> 841571 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:33109/solr/hdfsbackuprestore_testfailure_shard1_replica_t4/ shard1
   [junit4]   2> 841673 INFO  (zkCallback-5804-thread-3) [     ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/hdfsbackuprestore_testfailure/state.json] for collection [hdfsbackuprestore_testfailure] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 841674 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node11 x:hdfsbackuprestore_testfailure_shard2_replica_n8 ] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 841675 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr c:hdfsbackuprestore_testfailure s:shard1 r:core_node7 x:hdfsbackuprestore_testfailure_shard1_replica_t4 ] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 841679 INFO  (qtp2030067562-12292) [n:127.0.0.1:33109_solr     ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node11&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard2_replica_n8&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=921
   [junit4]   2> 841679 INFO  (qtp2030067562-12296) [n:127.0.0.1:33109_solr     ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard1_replica_t4&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard1&wt=javabin&version=2&replicaType=TLOG} status=0 QTime=923
   [junit4]   2> 841778 INFO  (zkCallback-5804-thread-3) [     ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/hdfsbackuprestore_testfailure/state.json] for collection [hdfsbackuprestore_testfailure] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 841778 INFO  (zkCallback-5804-thread-2) [     ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/hdfsbackuprestore_testfailure/state.json] for collection [hdfsbackuprestore_testfailure] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 842030 INFO  (qtp2030067562-12291) [n:127.0.0.1:33109_solr     ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard1_replica_n1&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1274
   [junit4]   2> 842044 INFO  (qtp805840179-12297) [n:127.0.0.1:32795_solr     ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node9&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testfailure_shard2_replica_n6&action=CREATE&numShards=2&collection=hdfsbackuprestore_testfailure&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1258
   [junit4]   2> 842047 INFO  (qtp805840179-12298) [n:127.0.0.1:32795_solr c:hdfsbackuprestore_testfailure s:shard2 r:core_node12 x:hdfsbackuprestore_testfailure_shard2_replica_t10 ] o.a.s.c.ZkController hdfsbackuprestore_test

[...truncated too long message...]

a:350) ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1275) [zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 863221 INFO  (closeThreadPool-5907-thread-1) [     ] o.a.s.c.Overseer Overseer (id=73836867949035526-127.0.0.1:32795_solr-n_0000000000) closing
   [junit4]   2> 863221 INFO  (zkCallback-5804-thread-2) [     ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
   [junit4]   2> 863224 INFO  (jetty-closer-5897-thread-2) [     ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.jvm tag=null
   [junit4]   2> 863224 INFO  (jetty-closer-5897-thread-2) [     ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@25c78f40: rootName = null, domain = solr.jvm, service url = null, agent id = null] for registry solr.jvm/com.codahale.metrics.MetricRegistry@3dc501aa
   [junit4]   2> 863227 INFO  (jetty-closer-5897-thread-2) [     ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.jetty tag=null
   [junit4]   2> 863227 INFO  (jetty-closer-5897-thread-2) [     ] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@3505ec2e: rootName = null, domain = solr.jetty, service url = null, agent id = null] for registry solr.jetty/com.codahale.metrics.MetricRegistry@4db9ba0a
   [junit4]   2> 863227 INFO  (jetty-closer-5897-thread-2) [     ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.cluster tag=null
   [junit4]   2> 863319 INFO  (jetty-closer-5897-thread-1) [     ] o.a.s.c.Overseer Overseer (id=73836867949035526-127.0.0.1:32795_solr-n_0000000000) closing
   [junit4]   2> 863320 INFO  (jetty-closer-5897-thread-1) [     ] o.e.j.s.AbstractConnector Stopped ServerConnector@3af04e43{SSL, (ssl, http/1.1)}{127.0.0.1:0}
   [junit4]   2> 863321 INFO  (jetty-closer-5897-thread-1) [     ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@357fa55f{/solr,null,STOPPED}
   [junit4]   2> 863321 INFO  (jetty-closer-5897-thread-1) [     ] o.e.j.s.session node0 Stopped scavenging
   [junit4]   2> 863422 INFO  (jetty-closer-5897-thread-2) [     ] o.e.j.s.AbstractConnector Stopped ServerConnector@73f8341a{SSL, (ssl, http/1.1)}{127.0.0.1:0}
   [junit4]   2> 863423 INFO  (jetty-closer-5897-thread-2) [     ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@3e102a18{/solr,null,STOPPED}
   [junit4]   2> 863423 INFO  (jetty-closer-5897-thread-2) [     ] o.e.j.s.session node0 Stopped scavenging
   [junit4]   2> 863424 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer Shutting down ZkTestServer.
   [junit4]   2> 863424 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker-SendThread(127.0.0.1:34839)) [     ] o.a.z.ClientCnxn An exception was thrown while closing send thread for session 0x106523d7d6e0000.
   [junit4]   2>           => EndOfStreamException: Unable to read additional data from server sessionid 0x106523d7d6e0000, likely server has closed socket
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77)
   [junit4]   2> org.apache.zookeeper.ClientCnxn$EndOfStreamException: Unable to read additional data from server sessionid 0x106523d7d6e0000, likely server has closed socket
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77) ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1275) [zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 863580 WARN  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker-SendThread(127.0.0.1:34839)) [     ] o.a.z.ClientCnxn An exception was thrown while closing send thread for session 0x106523d7d6e0001.
   [junit4]   2>           => EndOfStreamException: Unable to read additional data from server sessionid 0x106523d7d6e0001, likely server has closed socket
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77)
   [junit4]   2> org.apache.zookeeper.ClientCnxn$EndOfStreamException: Unable to read additional data from server sessionid 0x106523d7d6e0001, likely server has closed socket
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77) ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1275) [zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 863681 WARN  (ZkTestServer Run Thread) [     ] o.a.s.c.ZkTestServer Watch limit violations: 
   [junit4]   2> Maximum concurrent create/delete watches above limit:
   [junit4]   2> 
   [junit4]   2> 	9	/solr/collections/hdfsbackuprestore_testfailure/terms/shard2
   [junit4]   2> 	8	/solr/collections/hdfsbackuprestore_testfailure/terms/shard1
   [junit4]   2> 	7	/solr/collections/hdfsbackuprestore_testok/terms/shard1
   [junit4]   2> 	6	/solr/collections/hdfsbackuprestore_testok/terms/shard2
   [junit4]   2> 	3	/solr/collections/hdfsbackuprestore_testok_restored/terms/shard2
   [junit4]   2> 	3	/solr/collections/hdfsbackuprestore_testok_restored/terms/shard1
   [junit4]   2> 	3	/solr/aliases.json
   [junit4]   2> 	2	/solr/collections/hdfsbackuprestore_testok/collectionprops.json
   [junit4]   2> 	2	/solr/collections/hdfsbackuprestore_testok_restored/collectionprops.json
   [junit4]   2> 	2	/solr/packages.json
   [junit4]   2> 	2	/solr/security.json
   [junit4]   2> 	2	/solr/configs/customConfigName
   [junit4]   2> 	2	/solr/configs/conf1
   [junit4]   2> 	2	/solr/collections/hdfsbackuprestore_testfailure/collectionprops.json
   [junit4]   2> 	2	/solr/collections/hdfsbackuprestore_testfailure_restored/state.json
   [junit4]   2> 
   [junit4]   2> Maximum concurrent data watches above limit:
   [junit4]   2> 
   [junit4]   2> 	48	/solr/collections/hdfsbackuprestore_testok/state.json
   [junit4]   2> 	37	/solr/collections/hdfsbackuprestore_testok_restored/state.json
   [junit4]   2> 	24	/solr/collections/hdfsbackuprestore_testfailure/state.json
   [junit4]   2> 	3	/solr/clusterprops.json
   [junit4]   2> 	3	/solr/clusterstate.json
   [junit4]   2> 	3	/solr/collections/hdfsbackuprestore_testfailure_restored/state.json
   [junit4]   2> 
   [junit4]   2> Maximum concurrent children watches above limit:
   [junit4]   2> 
   [junit4]   2> 	91	/solr/overseer/queue
   [junit4]   2> 	18	/solr/collections
   [junit4]   2> 	18	/solr/overseer/collection-queue-work
   [junit4]   2> 	13	/solr/collections/hdfsbackuprestore_testfailure/state.json
   [junit4]   2> 	12	/solr/collections/hdfsbackuprestore_testok/state.json
   [junit4]   2> 	9	/solr/collections/hdfsbackuprestore_testok_restored/state.json
   [junit4]   2> 	7	/solr/live_nodes
   [junit4]   2> 	3	/solr/collections/hdfsbackuprestore_testfailure_restored/state.json
   [junit4]   2> 
   [junit4]   2> 863684 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer waitForServerDown: 127.0.0.1:34839
   [junit4]   2> 863684 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer parse host and port list: 127.0.0.1:34839
   [junit4]   2> 863684 INFO  (SUITE-TestHdfsCloudBackupRestore-seed#[1DF421E3E00928AB]-worker) [     ] o.a.s.c.ZkTestServer connecting to 127.0.0.1 34839
   [junit4]   2> NOTE: leaving temporary files on disk at: /home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.9/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_1DF421E3E00928AB-001
   [junit4]   2> Jul 12, 2021 3:23:16 PM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 16 leaked thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene87): {shard_s=PostingsFormat(name=LuceneFixedGap), _root_=PostingsFormat(name=LuceneFixedGap), id=PostingsFormat(name=LuceneVarGapFixedInterval)}, docValues:{}, maxPointsInLeafNode=153, maxMBSortInHeap=7.8590688783038685, sim=Asserting(RandomSimilarity(queryNorm=false): {}), locale=cs, timezone=Etc/Universal
   [junit4]   2> NOTE: Linux 5.4.0-45-generic amd64/Oracle Corporation 1.8.0_291 (64-bit)/cpus=4,threads=4,free=239288976,total=519569408
   [junit4]   2> NOTE: All tests run in this JVM: [TestContainerPlugin, TestSweetSpotSimilarityFactory, TestFieldCacheVsDocValues, RAMDirectoryFactoryTest, DeleteNodeTest, TestUseDocValuesAsStored, SolrCloudReportersTest, ChangedSchemaMergeTest, TestFacetMethods, TestDistributedMissingSort, TestCloudJSONFacetSKGEquiv, TestImpersonationWithHadoopAuth, TestSolr4Spatial2, BlobRepositoryCloudTest, ShardRoutingCustomTest, TestSimLargeCluster, TestPackages, HLLSerializationTest, TestCloudSearcherWarming, ZkControllerTest, TestRecoveryHdfs, DistribCursorPagingTest, ReplaceNodeTest, TestFieldCollectionResource, ZkCollectionPropsCachingTest, TestSkipOverseerOperations, TestMinHashQParser, TestSimDistributedQueue, TestCloudDeduplication, HdfsDirectoryTest, TestXIncludeConfig, TestCloudConsistency, CreateCollectionCleanupTest, PrimUtilsTest, TestSerializedLuceneMatchVersion, SignificantTermsQParserPluginTest, TestApiFramework, ZkSolrClientTest, ScheduledTriggerTest, PhrasesIdentificationComponentTest, ConcurrentCreateCollectionTest, TestChildDocTransformerHierarchy, TestManagedSynonymGraphFilterFactory, SolrTestCaseJ4Test, JsonLoaderTest, RouteFieldTest, TestNestedUpdateProcessor, TestConfigSetsAPIShareSchema, TestSQLHandler, BaseCdcrDistributedZkTest, TestSortByMinMaxFunction, TestHdfsCloudBackupRestore]
   [junit4] Completed [302/939 (1!)] on J0 in 36.17s, 2 tests, 1 failure <<< FAILURES!

[...truncated 54640 lines...]