You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2018/10/05 12:44:54 UTC
[JENKINS] Lucene-Solr-7.x-Windows (64bit/jdk-11) - Build # 820 -
Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Windows/820/
Java: 64bit/jdk-11 -XX:+UseCompressedOops -XX:+UseG1GC
1 tests failed.
FAILED: org.apache.solr.cloud.DocValuesNotIndexedTest.testGroupingDocAbsent
Error Message:
Did not find the expected number of groups for field boolGSF expected:<3> but was:<2>
Stack Trace:
java.lang.AssertionError: Did not find the expected number of groups for field boolGSF expected:<3> but was:<2>
at __randomizedtesting.SeedInfo.seed([493F376AA1E35A28:9788ECE244B21AED]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at org.junit.Assert.assertEquals(Assert.java:472)
at org.apache.solr.cloud.DocValuesNotIndexedTest.testGroupingDocAbsent(DocValuesNotIndexedTest.java:307)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:834)
Build Log:
[...truncated 13818 lines...]
[junit4] Suite: org.apache.solr.cloud.DocValuesNotIndexedTest
[junit4] 2> 1745005 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> Creating dataDir: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\init-core-data-001
[junit4] 2> 1745006 WARN (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=4 numCloses=4
[junit4] 2> 1745006 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 1745008 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
[junit4] 2> 1745008 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001
[junit4] 2> 1745009 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1745009 INFO (Thread-2956) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1745009 INFO (Thread-2956) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 1745013 ERROR (Thread-2956) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 1745110 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.c.ZkTestServer start zk server on port:61049
[junit4] 2> 1745113 INFO (zkConnectionManagerCallback-4344-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745113 INFO (jetty-launcher-4341-thread-2) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 11+28
[junit4] 2> 1745118 INFO (jetty-launcher-4341-thread-2) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1745118 INFO (jetty-launcher-4341-thread-2) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1745118 INFO (jetty-launcher-4341-thread-2) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1745118 INFO (jetty-launcher-4341-thread-1) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 11+28
[junit4] 2> 1745118 INFO (jetty-launcher-4341-thread-2) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@6be1bd46{/solr,null,AVAILABLE}
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@26adb66e{/solr,null,AVAILABLE}
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.e.j.s.AbstractConnector Started ServerConnector@76aefa6a{HTTP/1.1,[http/1.1]}{127.0.0.1:61053}
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.e.j.s.Server Started @1745154ms
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=61053}
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@5981e45c{HTTP/1.1,[http/1.1]}{127.0.0.1:61056}
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.e.j.s.Server Started @1745155ms
[junit4] 2> 1745119 ERROR (jetty-launcher-4341-thread-2) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=61056}
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.6.0
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-10-05T11:27:30.924757400Z
[junit4] 2> 1745119 ERROR (jetty-launcher-4341-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.6.0
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1745119 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-10-05T11:27:30.924757400Z
[junit4] 2> 1745125 INFO (zkConnectionManagerCallback-4348-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745126 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1745126 INFO (zkConnectionManagerCallback-4346-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745126 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1745311 INFO (jetty-launcher-4341-thread-2) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:61049/solr
[junit4] 2> 1745315 INFO (zkConnectionManagerCallback-4352-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745318 INFO (zkConnectionManagerCallback-4354-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745351 INFO (jetty-launcher-4341-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:61049/solr
[junit4] 2> 1745353 INFO (zkConnectionManagerCallback-4360-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745359 INFO (zkConnectionManagerCallback-4362-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745373 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:61056_solr
[junit4] 2> 1745373 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.c.Overseer Overseer (id=72169633326825478-127.0.0.1:61056_solr-n_0000000000) starting
[junit4] 2> 1745386 INFO (zkConnectionManagerCallback-4369-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745387 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:61049/solr ready
[junit4] 2> 1745389 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:61056_solr
[junit4] 2> 1745417 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1745425 INFO (OverseerStateUpdate-72169633326825478-127.0.0.1:61056_solr-n_0000000000) [n:127.0.0.1:61056_solr ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:61056_solr
[junit4] 2> 1745434 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1745438 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61056.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1745438 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1745438 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:61053_solr
[junit4] 2> 1745447 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61056.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1745447 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61056.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1745449 INFO (jetty-launcher-4341-thread-1) [n:127.0.0.1:61056_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\.
[junit4] 2> 1745468 INFO (OverseerStateUpdate-72169633326825478-127.0.0.1:61056_solr-n_0000000000) [n:127.0.0.1:61056_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1745473 INFO (zkCallback-4368-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1745474 INFO (zkCallback-4353-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1745478 INFO (zkConnectionManagerCallback-4375-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745479 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1745480 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:61049/solr ready
[junit4] 2> 1745480 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1745501 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61053.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1745511 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61053.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1745511 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61053.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1745514 INFO (jetty-launcher-4341-thread-2) [n:127.0.0.1:61053_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node2\.
[junit4] 2> 1745536 INFO (zkConnectionManagerCallback-4378-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745538 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x10065e62e040009, likely client has closed socket
[junit4] 2> 1745541 INFO (zkConnectionManagerCallback-4383-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1745542 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1745544 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:61049/solr ready
[junit4] 2> 1745551 INFO (qtp415149573-17941) [n:127.0.0.1:61053_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&maxShardsPerNode=2&name=dv_coll&nrtReplicas=1&action=CREATE&numShards=4&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1745553 INFO (OverseerThreadFactory-7836-thread-1-processing-n:127.0.0.1:61056_solr) [n:127.0.0.1:61056_solr ] o.a.s.c.a.c.CreateCollectionCmd Create collection dv_coll
[junit4] 2> 1745663 INFO (OverseerStateUpdate-72169633326825478-127.0.0.1:61056_solr-n_0000000000) [n:127.0.0.1:61056_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"dv_coll",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"dv_coll_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:61056/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1745664 INFO (OverseerStateUpdate-72169633326825478-127.0.0.1:61056_solr-n_0000000000) [n:127.0.0.1:61056_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"dv_coll",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"dv_coll_shard2_replica_n2",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:61053/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1745666 INFO (OverseerStateUpdate-72169633326825478-127.0.0.1:61056_solr-n_0000000000) [n:127.0.0.1:61056_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"dv_coll",
[junit4] 2> "shard":"shard3",
[junit4] 2> "core":"dv_coll_shard3_replica_n4",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:61056/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1745668 INFO (OverseerStateUpdate-72169633326825478-127.0.0.1:61056_solr-n_0000000000) [n:127.0.0.1:61056_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"dv_coll",
[junit4] 2> "shard":"shard4",
[junit4] 2> "core":"dv_coll_shard4_replica_n6",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:61053/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1745872 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr x:dv_coll_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=dv_coll_shard1_replica_n1&action=CREATE&numShards=4&collection=dv_coll&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1745872 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr x:dv_coll_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1745883 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr x:dv_coll_shard2_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=dv_coll_shard2_replica_n2&action=CREATE&numShards=4&collection=dv_coll&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1745888 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr x:dv_coll_shard4_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=dv_coll_shard4_replica_n6&action=CREATE&numShards=4&collection=dv_coll&shard=shard4&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1745889 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr x:dv_coll_shard3_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=dv_coll_shard3_replica_n4&action=CREATE&numShards=4&collection=dv_coll&shard=shard3&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1746890 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1746901 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1746903 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1746903 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll, trusted=true
[junit4] 2> 1746904 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61056.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1746904 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\dv_coll_shard1_replica_n1], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\.\dv_coll_shard1_replica_n1\data\]
[junit4] 2> 1746910 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1746912 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1746914 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1746918 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema [dv_coll_shard3_replica_n4] Schema name=minimal
[junit4] 2> 1746921 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.IndexSchema [dv_coll_shard2_replica_n2] Schema name=minimal
[junit4] 2> 1746921 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1746921 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard3_replica_n4' using configuration from collection dv_coll, trusted=true
[junit4] 2> 1746922 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61056.solr.core.dv_coll.shard3.replica_n4' (registry 'solr.core.dv_coll.shard3.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1746922 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [[dv_coll_shard3_replica_n4] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\dv_coll_shard3_replica_n4], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\.\dv_coll_shard3_replica_n4\data\]
[junit4] 2> 1746926 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1746926 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard2_replica_n2' using configuration from collection dv_coll, trusted=true
[junit4] 2> 1746927 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61053.solr.core.dv_coll.shard2.replica_n2' (registry 'solr.core.dv_coll.shard2.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1746927 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [[dv_coll_shard2_replica_n2] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node2\dv_coll_shard2_replica_n2], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node2\.\dv_coll_shard2_replica_n2\data\]
[junit4] 2> 1746932 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.IndexSchema [dv_coll_shard4_replica_n6] Schema name=minimal
[junit4] 2> 1746933 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1746933 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard4_replica_n6' using configuration from collection dv_coll, trusted=true
[junit4] 2> 1746934 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61053.solr.core.dv_coll.shard4.replica_n6' (registry 'solr.core.dv_coll.shard4.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1746934 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [[dv_coll_shard4_replica_n6] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node2\dv_coll_shard4_replica_n6], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node2\.\dv_coll_shard4_replica_n6\data\]
[junit4] 2> 1747068 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1747068 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1747070 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1747070 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1747078 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1747078 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1747080 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1747080 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1747088 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1747088 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1747090 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1747090 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1747091 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@163a48c4[dv_coll_shard1_replica_n1] main]
[junit4] 2> 1747092 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1747092 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@3efd3007[dv_coll_shard2_replica_n2] main]
[junit4] 2> 1747092 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1747092 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
[junit4] 2> 1747093 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 1747093 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1747093 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613484631416373248
[junit4] 2> 1747094 INFO (searcherExecutor-7840-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard1_replica_n1 c:dv_coll s:shard1 r:core_node3) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@163a48c4[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1747094 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1747095 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1747095 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
[junit4] 2> 1747095 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 1747095 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1747095 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613484631418470400
[junit4] 2> 1747095 INFO (searcherExecutor-7842-thread-1-processing-n:127.0.0.1:61053_solr x:dv_coll_shard2_replica_n2 c:dv_coll s:shard2 r:core_node5) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [dv_coll_shard2_replica_n2] Registered new searcher Searcher@3efd3007[dv_coll_shard2_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1747098 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@243feb8f[dv_coll_shard4_replica_n6] main]
[junit4] 2> 1747098 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard1 to Terms{values={core_node3=0}, version=0}
[junit4] 2> 1747100 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1747101 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1747101 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
[junit4] 2> 1747101 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 1747101 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard2 to Terms{values={core_node5=0}, version=0}
[junit4] 2> 1747101 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1747101 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1747101 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:61056/solr/dv_coll_shard1_replica_n1/
[junit4] 2> 1747102 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1747102 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1747103 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy http://127.0.0.1:61056/solr/dv_coll_shard1_replica_n1/ has no replicas
[junit4] 2> 1747103 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1747103 INFO (searcherExecutor-7843-thread-1-processing-n:127.0.0.1:61053_solr x:dv_coll_shard4_replica_n6 c:dv_coll s:shard4 r:core_node8) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6] Registered new searcher Searcher@243feb8f[dv_coll_shard4_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1747103 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613484631426859008
[junit4] 2> 1747104 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1747105 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1747105 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:61053/solr/dv_coll_shard2_replica_n2/
[junit4] 2> 1747105 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1747105 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy http://127.0.0.1:61053/solr/dv_coll_shard2_replica_n2/ has no replicas
[junit4] 2> 1747105 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1747105 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:61056/solr/dv_coll_shard1_replica_n1/ shard1
[junit4] 2> 1747108 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:61053/solr/dv_coll_shard2_replica_n2/ shard2
[junit4] 2> 1747109 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard4 to Terms{values={core_node8=0}, version=0}
[junit4] 2> 1747110 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1747110 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1747112 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1747112 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1747112 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:61053/solr/dv_coll_shard4_replica_n6/
[junit4] 2> 1747112 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1747112 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy http://127.0.0.1:61053/solr/dv_coll_shard4_replica_n6/ has no replicas
[junit4] 2> 1747112 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1747113 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1747113 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1747114 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:61053/solr/dv_coll_shard4_replica_n6/ shard4
[junit4] 2> 1747116 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@2caf50b3[dv_coll_shard3_replica_n4] main]
[junit4] 2> 1747116 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1747116 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1747116 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
[junit4] 2> 1747117 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 1747117 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1747118 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613484631442587648
[junit4] 2> 1747118 INFO (searcherExecutor-7841-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard3_replica_n4 c:dv_coll s:shard3 r:core_node7) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] Registered new searcher Searcher@2caf50b3[dv_coll_shard3_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1747121 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard3 to Terms{values={core_node7=0}, version=0}
[junit4] 2> 1747124 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1747124 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1747124 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/
[junit4] 2> 1747124 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1747124 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/ has no replicas
[junit4] 2> 1747124 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1747128 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/ shard3
[junit4] 2> 1747261 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1747261 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1747263 INFO (qtp1813473807-17947) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=dv_coll_shard1_replica_n1&action=CREATE&numShards=4&collection=dv_coll&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1392
[junit4] 2> 1747264 INFO (qtp415149573-17936) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=dv_coll_shard2_replica_n2&action=CREATE&numShards=4&collection=dv_coll&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1380
[junit4] 2> 1747267 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1747270 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=dv_coll_shard4_replica_n6&action=CREATE&numShards=4&collection=dv_coll&shard=shard4&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1381
[junit4] 2> 1747280 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1747283 INFO (qtp1813473807-17952) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=dv_coll_shard3_replica_n4&action=CREATE&numShards=4&collection=dv_coll&shard=shard3&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1393
[junit4] 2> 1747285 INFO (qtp415149573-17941) [n:127.0.0.1:61053_solr ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
[junit4] 2> 1747384 INFO (zkCallback-4353-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1747384 INFO (zkCallback-4361-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1747555 INFO (OverseerCollectionConfigSetProcessor-72169633326825478-127.0.0.1:61056_solr-n_0000000000) [n:127.0.0.1:61056_solr ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 1748286 INFO (qtp415149573-17941) [n:127.0.0.1:61053_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&maxShardsPerNode=2&name=dv_coll&nrtReplicas=1&action=CREATE&numShards=4&wt=javabin&version=2} status=0 QTime=2735
[junit4] 2> 1748292 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SchemaManager managed schema loaded . version : 0
[junit4] 2> 1748294 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1748296 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1748304 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkController Persisted config data to node /configs/conf1/managed-schema
[junit4] 2> 1748304 INFO (zkCallback-4361-thread-2) [ ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
[junit4] 2> 1748304 INFO (zkCallback-4353-thread-2) [ ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
[junit4] 2> 1748304 INFO (zkCallback-4353-thread-1) [ ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
[junit4] 2> 1748304 INFO (zkCallback-4361-thread-1) [ ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
[junit4] 2> 1748305 INFO (zkCallback-4353-thread-2) [ ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
[junit4] 2> 1748305 INFO (zkCallback-4353-thread-1) [ ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
[junit4] 2> 1748306 INFO (zkCallback-4353-thread-2) [ ] o.a.s.s.IndexSchema [dv_coll_shard2_replica_n2] Schema name=minimal
[junit4] 2> 1748306 INFO (zkCallback-4361-thread-1) [ ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
[junit4] 2> 1748306 INFO (zkCallback-4353-thread-1) [ ] o.a.s.s.IndexSchema [dv_coll_shard4_replica_n6] Schema name=minimal
[junit4] 2> 1748306 INFO (zkCallback-4361-thread-1) [ ] o.a.s.s.IndexSchema [dv_coll_shard3_replica_n4] Schema name=minimal
[junit4] 2> 1748326 INFO (zkCallback-4353-thread-1) [ ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1748326 INFO (zkCallback-4353-thread-1) [ ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 20 ms
[junit4] 2> 1748329 INFO (zkCallback-4353-thread-2) [ ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1748330 INFO (zkCallback-4361-thread-1) [ ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1748330 INFO (zkCallback-4353-thread-2) [ ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 25 ms
[junit4] 2> 1748331 INFO (zkCallback-4361-thread-1) [ ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 23 ms
[junit4] 2> 1748337 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1748343 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1748346 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1748346 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll
[junit4] 2> 1748348 INFO (Thread-2980) [ ] o.a.s.c.SolrCore config update listener called for core dv_coll_shard2_replica_n2
[junit4] 2> 1748368 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61056.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1748368 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\dv_coll_shard1_replica_n1], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\.\dv_coll_shard1_replica_n1\data\]
[junit4] 2> 1748370 INFO (Thread-2979) [ ] o.a.s.c.SolrCore config update listener called for core dv_coll_shard1_replica_n1
[junit4] 2> 1748373 INFO (Thread-2980) [ ] o.a.s.c.SolrCore config update listener called for core dv_coll_shard4_replica_n6
[junit4] 2> 1748373 INFO (Thread-2979) [ ] o.a.s.c.SolrCore core reload dv_coll_shard1_replica_n1
[junit4] 2> 1748490 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1748490 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1748492 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@c7056ed[dv_coll_shard1_replica_n1] main]
[junit4] 2> 1748493 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1748494 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1748494 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
[junit4] 2> 1748494 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 1 is already the latest
[junit4] 2> 1748494 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1748495 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613484632886476800
[junit4] 2> 1748495 INFO (searcherExecutor-7860-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard1_replica_n1 c:dv_coll s:shard1 r:core_node3) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@c7056ed[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1748496 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DefaultSolrCoreState New IndexWriter is ready to be used.
[junit4] 2> 1748497 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@3eea7c68[dv_coll_shard1_replica_n1] main]
[junit4] 2> 1748498 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ManagedIndexSchema Waiting up to 599 secs for 3 replicas to apply schema update version 1 for collection dv_coll
[junit4] 2> 1748498 INFO (searcherExecutor-7860-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard1_replica_n1 c:dv_coll s:shard1 r:core_node3) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@3eea7c68[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1748499 INFO (zkCallback-4361-thread-2) [ ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
[junit4] 2> 1748501 INFO (zkCallback-4361-thread-2) [ ] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1748503 INFO (zkCallback-4361-thread-2) [ ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1748503 INFO (zkCallback-4361-thread-2) [ ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 4 ms
[junit4] 2> 1748505 INFO (Thread-2979) [ ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1748509 INFO (Thread-2979) [ ] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1748512 INFO (Thread-2979) [ ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1748512 INFO (Thread-2979) [ ] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll
[junit4] 2> 1748526 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_61056.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5c9d7e78
[junit4] 2> 1748526 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\dv_coll_shard1_replica_n1], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001\tempDir-001\node1\.\dv_coll_shard1_replica_n1\data\]
[junit4] 2> 1748533 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.S.Request [dv_coll_shard2_replica_n2] webapp=/solr path=/schema/zkversion params={refreshIfBelowVersion=1&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 1748535 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.S.Request [dv_coll_shard3_replica_n4] webapp=/solr path=/schema/zkversion params={refreshIfBelowVersion=1&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 1748537 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.S.Request [dv_coll_shard4_replica_n6] webapp=/solr path=/schema/zkversion params={refreshIfBelowVersion=1&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 1748538 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ManagedIndexSchema Took 40.0ms for 3 replicas to apply schema update version 1 for collection dv_coll
[junit4] 2> 1748538 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.S.Request [dv_coll_shard1_replica_n1] webapp=/solr path=/schema params={wt=javabin&version=2} status=0 QTime=246
[junit4] 2> 1748538 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] CLOSING SolrCore org.apache.solr.core.SolrCore@64fba6bd
[junit4] 2> 1748538 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard1.replica_n1, tag=64fba6bd
[junit4] 2> 1748538 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@3da3deb5: rootName = solr_61056, domain = solr.core.dv_coll.shard1.replica_n1, service url = null, agent id = null] for registry solr.core.dv_coll.shard1.replica_n1 / com.codahale.metrics.MetricRegistry@3711f608
[junit4] 2> 1748544 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard1.leader, tag=64fba6bd
[junit4] 2> 1748574 INFO (TEST-DocValuesNotIndexedTest.testGroupingDocAbsent-seed#[493F376AA1E35A28]) [ ] o.a.s.SolrTestCaseJ4 ###Starting testGroupingDocAbsent
[junit4] 2> 1748581 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard3 to Terms{values={core_node7=1}, version=1}
[junit4] 2> 1748586 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard4 to Terms{values={core_node8=1}, version=1}
[junit4] 2> 1748586 INFO (qtp415149573-17942) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard4_replica_n6] webapp=/solr path=/update params={update.distrib=TOLEADER&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&wt=javabin&version=2}{deleteByQuery=*:* (-1613484632977702912)} 0 4
[junit4] 2> 1748587 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard1 to Terms{values={core_node3=1}, version=1}
[junit4] 2> 1748587 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard1_replica_n1] webapp=/solr path=/update params={update.distrib=TOLEADER&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&wt=javabin&version=2}{deleteByQuery=*:* (-1613484632977702912)} 0 5
[junit4] 2> 1748643 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/dv_coll/terms/shard2 to Terms{values={core_node5=1}, version=1}
[junit4] 2> 1748643 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard2_replica_n2] webapp=/solr path=/update params={update.distrib=TOLEADER&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&wt=javabin&version=2}{deleteByQuery=*:* (-1613484632975605760)} 0 63
[junit4] 2> 1748644 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4] webapp=/solr path=/update params={wt=javabin&version=2}{deleteByQuery=*:* (-1613484632970362880)} 0 69
[junit4] 2> 1748645 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633043763200,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748645 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1748646 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633044811776,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748646 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1748647 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@1b1bb4e4[dv_coll_shard2_replica_n2] main]
[junit4] 2> 1748647 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748649 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633047957504,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748649 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1748649 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@6e63418e[dv_coll_shard3_replica_n4] main]
[junit4] 2> 1748649 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748649 INFO (searcherExecutor-7842-thread-1-processing-n:127.0.0.1:61053_solr x:dv_coll_shard2_replica_n2 c:dv_coll s:shard2 r:core_node5) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [dv_coll_shard2_replica_n2] Registered new searcher Searcher@1b1bb4e4[dv_coll_shard2_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1748650 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard2_replica_n2] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 3
[junit4] 2> 1748650 INFO (searcherExecutor-7841-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard3_replica_n4 c:dv_coll s:shard3 r:core_node7) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] Registered new searcher Searcher@6e63418e[dv_coll_shard3_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1748650 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 2
[junit4] 2> 1748651 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748651 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard1_replica_n1] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 5
[junit4] 2> 1748652 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633051103232,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748652 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1748652 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@41a59e79[dv_coll_shard4_replica_n6] main]
[junit4] 2> 1748653 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748654 INFO (searcherExecutor-7843-thread-1-processing-n:127.0.0.1:61053_solr x:dv_coll_shard4_replica_n6 c:dv_coll s:shard4 r:core_node8) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6] Registered new searcher Searcher@41a59e79[dv_coll_shard4_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1748654 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard4_replica_n6] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 2
[junit4] 2> 1748654 INFO (qtp1813473807-17960) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4] webapp=/solr path=/update params={_stateVer_=dv_coll:4&waitSearcher=true&commit=true&softCommit=false&wt=javabin&version=2}{commit=} 0 9
[junit4] 2> 1748659 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard2_replica_n2] webapp=/solr path=/update params={_stateVer_=dv_coll:4&wt=javabin&version=2}{add=[4 (1613484633053200384)]} 0 2
[junit4] 2> 1748660 INFO (qtp1813473807-17949) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4] webapp=/solr path=/update params={_stateVer_=dv_coll:4&wt=javabin&version=2}{add=[2 (1613484633053200384), 3 (1613484633059491840)]} 0 3
[junit4] 2> 1748665 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1748665 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1748666 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@548aca70[dv_coll_shard1_replica_n1] main]
[junit4] 2> 1748668 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard1_replica_n1] webapp=/solr path=/update params={_stateVer_=dv_coll:4&wt=javabin&version=2}{add=[1 (1613484633066831872)]} 0 1
[junit4] 2> 1748670 INFO (qtp1813473807-17949) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633069977600,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748670 INFO (qtp1813473807-17949) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@b3e1f18 commitCommandVersion:1613484633069977600
[junit4] 2> 1748670 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633069977600,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748670 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@3d7e7839 commitCommandVersion:1613484633069977600
[junit4] 2> 1748671 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633071026176,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748671 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@2fb8d028 commitCommandVersion:1613484633071026176
[junit4] 2> 1748672 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1613484633072074752,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1748672 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1748672 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 1748673 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1748673 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
[junit4] 2> 1748673 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 1 is already the latest
[junit4] 2> 1748674 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1748674 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748674 INFO (qtp415149573-17938) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard4_replica_n6] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 2
[junit4] 2> 1748676 INFO (searcherExecutor-7865-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard1_replica_n1 c:dv_coll s:shard1 r:core_node3) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@548aca70[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1748726 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@4edc4581[dv_coll_shard2_replica_n2] main]
[junit4] 2> 1748726 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748727 INFO (searcherExecutor-7842-thread-1-processing-n:127.0.0.1:61053_solr x:dv_coll_shard2_replica_n2 c:dv_coll s:shard2 r:core_node5) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [dv_coll_shard2_replica_n2] Registered new searcher Searcher@4edc4581[dv_coll_shard2_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(7.6.0):C1)))}
[junit4] 2> 1748727 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard2_replica_n2] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 57
[junit4] 2> 1748728 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@3500ae8e[dv_coll_shard3_replica_n4] main]
[junit4] 2> 1748729 INFO (searcherExecutor-7841-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard3_replica_n4 c:dv_coll s:shard3 r:core_node7) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] Registered new searcher Searcher@3500ae8e[dv_coll_shard3_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(7.6.0):C2)))}
[junit4] 2> 1748729 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748729 INFO (qtp1813473807-17948) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 57
[junit4] 2> 1748730 INFO (qtp1813473807-17949) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@31e5c668[dv_coll_shard1_replica_n1] main]
[junit4] 2> 1748731 INFO (qtp1813473807-17949) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1748731 INFO (searcherExecutor-7860-thread-1-processing-n:127.0.0.1:61056_solr x:dv_coll_shard1_replica_n1 c:dv_coll s:shard1 r:core_node3) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@31e5c668[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(7.6.0):C1)))}
[junit4] 2> 1748731 INFO (qtp1813473807-17949) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard1_replica_n1] webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 61
[junit4] 2> 1748732 INFO (qtp1813473807-17945) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4] webapp=/solr path=/update params={_stateVer_=dv_coll:4&commit=true&wt=javabin&version=2}{commit=} 0 62
[junit4] 2> 1748736 INFO (Thread-2979) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DefaultSolrCoreState New IndexWriter is ready to be used.
[junit4] 2> 1748750 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.S.Request [dv_coll_shard2_replica_n2] webapp=/solr path=/select params={df=text&group.distributed.first=true&distrib=false&_stateVer_=dv_coll:4&fl=id,score&shards.purpose=2048&start=0&shard.url=http://127.0.0.1:61053/solr/dv_coll_shard2_replica_n2/&rows=10&version=2&q=*:*&NOW=1538738854537&isShard=true&wt=javabin&group.field=intGSF&group=true} status=0 QTime=12
[junit4] 2> 1748754 INFO (qtp415149573-17940) [n:127.0.0.1:61053_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.S.Request [dv_coll_shard4_replica_n6] webapp=/solr path=/select params={df=text&group.distributed.first=true&distrib=false&_stateVer_=dv_coll:4&fl=id,score&shards.purpose=2048&start=0&shard.url=http://127.0.0.1:61053/solr/dv_coll_shard4_replica_n6/&rows=10&version=2&q=*:*&NOW=1538738854537&isShard=true&wt=javabin&group.field=intGSF&group=true} status=0 QTime=0
[junit4] 2> 1748754 INFO (qtp1813473807-17951) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.S.Request [dv_coll_shard3_replica_n4] webapp=/solr path=/select params={df=text&group.distributed.first=true&distrib=false&_stateVer_=dv_coll:4&fl=id,score&shards.purpose=2048&start=0&shard.url=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&rows=10&version=2&q=*:*&NOW=1538738854537&isShard=true&wt=javabin&group.field=intGSF&group=true} status=0 QTime=0
[junit4] 2> 1748769 INFO (qtp1813473807-17951) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.S.Request [dv_coll_shard1_replica_n1] webapp=/solr path=/select params={df=text&group.distributed.first=true&distrib=false&_stateVer_=dv_coll:4&fl=id,score&shards.purpose=2048&start=0&shard.url=http://127.0.0.1:61056/solr/dv_coll_shard1_replica_n1/&rows=10&version=2&q=*:*&NOW=1538738854537&isShard=true&wt=javabin&group.field=intGSF&group=true} status=0 QTime=0
[junit4] 2> 1748794 INFO (qtp1813473807-17951) [n:127.0.0.1:61056_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.S.Request [dv_coll_shard3_replica_n4] webapp=/solr path=/select params={df=text&distrib=false&_stateVer_=dv_coll:4&fl=id,score&shards.purpose=4&start=0&shard.url=http://127.0.0.1:61056/solr/dv_coll_shard3_replica_n4/&rows=10&version=2&group.distributed.second=true&q=*:*&NOW=1538738854537&group.topgroups.intGSF=%01&group.topgroups.intGSF=805339320&group.topgroups.intGSF=805359325&group.topgroups.intGSF=805349322&isShard=true&wt=javabin&group.field=intGSF&group=true} status=0 QTime=17
[junit4] 2> 1748796 INFO (qtp1813473807-17951) [n:127.0.0.1:61056_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.S.Request [dv_coll_shard1_replica_n1] webapp=/solr path=/select params={df=text&distrib=false&_stateVer_=dv_coll:4&fl=id,score&shards.purpose=4&start=0&shard.url=http://127.0
[...truncated too long message...]
unit4] 2> 25660 INFO (coreCloseExecutor-58-thread-2) [n:127.0.0.1:56679_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@67a61aa0: rootName = solr_56679, domain = solr.core.dv_coll.shard1.replica_n1, service url = null, agent id = null] for registry solr.core.dv_coll.shard1.replica_n1 / com.codahale.metrics.MetricRegistry@5efef108
[junit4] 2> 25668 INFO (coreCloseExecutor-58-thread-2) [n:127.0.0.1:56679_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard1.leader, tag=13b80593
[junit4] 2> 25670 INFO (coreCloseExecutor-58-thread-1) [n:127.0.0.1:56679_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] CLOSING SolrCore org.apache.solr.core.SolrCore@77256799
[junit4] 2> 25679 INFO (coreCloseExecutor-58-thread-1) [n:127.0.0.1:56679_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard3.replica_n4, tag=77256799
[junit4] 2> 25679 INFO (coreCloseExecutor-58-thread-1) [n:127.0.0.1:56679_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@5b21118a: rootName = solr_56679, domain = solr.core.dv_coll.shard3.replica_n4, service url = null, agent id = null] for registry solr.core.dv_coll.shard3.replica_n4 / com.codahale.metrics.MetricRegistry@79a25746
[junit4] 2> 25692 INFO (coreCloseExecutor-58-thread-1) [n:127.0.0.1:56679_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard3.leader, tag=77256799
[junit4] 2> 25694 INFO (zkCallback-19-thread-2) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [0])
[junit4] 2> 25697 INFO (zkCallback-17-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [0])
[junit4] 2> 25715 INFO (jetty-closer-2-thread-2) [ ] o.a.s.c.Overseer Overseer (id=72169929971662854-127.0.0.1:56679_solr-n_0000000000) closing
[junit4] 2> 25715 INFO (OverseerStateUpdate-72169929971662854-127.0.0.1:56679_solr-n_0000000000) [n:127.0.0.1:56679_solr ] o.a.s.c.Overseer Overseer Loop exiting : 127.0.0.1:56679_solr
[junit4] 2> 25720 WARN (OverseerAutoScalingTriggerThread-72169929971662854-127.0.0.1:56679_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
[junit4] 2> 25722 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100662b3f6d0007, likely client has closed socket
[junit4] 2> 25724 INFO (jetty-closer-2-thread-1) [ ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@6dca9d5{/solr,null,UNAVAILABLE}
[junit4] 2> 25724 INFO (jetty-closer-2-thread-1) [ ] o.e.j.s.session node0 Stopped scavenging
[junit4] 2> 25726 INFO (jetty-closer-2-thread-2) [ ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@5ad01052{/solr,null,UNAVAILABLE}
[junit4] 2> 25727 INFO (jetty-closer-2-thread-2) [ ] o.e.j.s.session node0 Stopped scavenging
[junit4] 2> 25728 ERROR (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 25729 INFO (SUITE-DocValuesNotIndexedTest-seed#[493F376AA1E35A28]-worker) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:56672 56672
[junit4] 2> 25730 INFO (Thread-0) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:56672 56672
[junit4] 2> 25731 WARN (Thread-0) [ ] o.a.s.c.ZkTestServer Watch limit violations:
[junit4] 2> Maximum concurrent create/delete watches above limit:
[junit4] 2>
[junit4] 2> 7 /solr/configs/conf1/managed-schema
[junit4] 2> 5 /solr/aliases.json
[junit4] 2> 5 /solr/clusterprops.json
[junit4] 2> 4 /solr/configs/conf1
[junit4] 2> 3 /solr/collections/dv_coll/terms/shard3
[junit4] 2> 3 /solr/collections/dv_coll/terms/shard2
[junit4] 2> 3 /solr/collections/dv_coll/terms/shard1
[junit4] 2> 3 /solr/collections/dv_coll/terms/shard4
[junit4] 2> 2 /solr/security.json
[junit4] 2>
[junit4] 2> Maximum concurrent data watches above limit:
[junit4] 2>
[junit4] 2> 6 /solr/collections/dv_coll/state.json
[junit4] 2> 5 /solr/clusterstate.json
[junit4] 2> 2 /solr/configs/conf1/managed-schema
[junit4] 2>
[junit4] 2> Maximum concurrent children watches above limit:
[junit4] 2>
[junit4] 2> 5 /solr/live_nodes
[junit4] 2> 5 /solr/collections
[junit4] 2>
[junit4] 2> NOTE: leaving temporary files on disk at: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J1\temp\solr.cloud.DocValuesNotIndexedTest_493F376AA1E35A28-001
[junit4] 2> Oct 05, 2018 12:43:11 PM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 5 leaked thread(s).
[junit4] 2> NOTE: test params are: codec=Asserting(Lucene70): {id=Lucene50(blocksize=128)}, docValues:{floatField=DocValuesFormat(name=Asserting), doubleField=DocValuesFormat(name=Direct), intFieldMulti=DocValuesFormat(name=Asserting), boolFieldMulti=DocValuesFormat(name=Direct), floatGSL=DocValuesFormat(name=Memory), longFieldMulti=DocValuesFormat(name=Lucene70), intField=DocValuesFormat(name=Direct), boolField=DocValuesFormat(name=Memory), stringFieldMulti=DocValuesFormat(name=Memory), dateFieldMulti=DocValuesFormat(name=Lucene70), longGSF=DocValuesFormat(name=Asserting), doubleFieldMulti=DocValuesFormat(name=Memory), stringGSL=DocValuesFormat(name=Lucene70), doubleGSF=DocValuesFormat(name=Direct), doubleGSL=DocValuesFormat(name=Lucene70), stringGSF=DocValuesFormat(name=Direct), intGSL=DocValuesFormat(name=Direct), longField=DocValuesFormat(name=Asserting), dateField=DocValuesFormat(name=Memory), dateGSF=DocValuesFormat(name=Memory), boolGSF=DocValuesFormat(name=Memory), longGSL=DocValuesFormat(name=Memory), boolGSL=DocValuesFormat(name=Asserting), floatFieldMulti=DocValuesFormat(name=Direct), floatGSF=DocValuesFormat(name=Asserting), dateGSL=DocValuesFormat(name=Asserting), _version_=DocValuesFormat(name=Asserting), intGSF=DocValuesFormat(name=Lucene70), stringField=DocValuesFormat(name=Direct)}, maxPointsInLeafNode=617, maxMBSortInHeap=5.600807087073306, sim=RandomSimilarity(queryNorm=false): {}, locale=rw-RW, timezone=America/Juneau
[junit4] 2> NOTE: Windows 10 10.0 amd64/Oracle Corporation 11 (64-bit)/cpus=3,threads=1,free=39991352,total=100663296
[junit4] 2> NOTE: All tests run in this JVM: [DocValuesNotIndexedTest]
[junit4] Completed [1/5 (1!)] on J1 in 24.74s, 4 tests, 1 failure <<< FAILURES!
[...truncated 39 lines...]
BUILD FAILED
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\common-build.xml:1567: The following error occurred while executing this line:
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\common-build.xml:1092: There were test failures: 5 suites, 20 tests, 1 failure [seed: 493F376AA1E35A28]
Total time: 41 seconds
[repro] Setting last failure code to 1
[repro] Failures:
[repro] 1/5 failed: org.apache.solr.cloud.DocValuesNotIndexedTest
[repro] Exiting with code 1
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows>IF EXIST lucene\build move /y lucene\build lucene\build.repro
1 dir(s) moved.
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows>IF EXIST solr\build move /y solr\build solr\build.repro
1 dir(s) moved.
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows>IF EXIST lucene\build.orig move /y lucene\build.orig lucene\build
1 dir(s) moved.
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows>IF EXIST solr\build.orig move /y solr\build.orig solr\build
1 dir(s) moved.
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows>exit 0
Archiving artifacts
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
[WARNINGS] Parsing warnings in console log with parser Java Compiler (javac)
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=ab50275baea80b3fc41a5054708ec490cab0fa06, workspace=C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows
[WARNINGS] Computing warning deltas based on reference build #819
Recording test results
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
[JENKINS] Lucene-Solr-7.x-Windows (64bit/jdk-11) - Build # 821 -
Failure!
Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Windows/821/
Java: 64bit/jdk-11 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC
2 tests failed.
FAILED: junit.framework.TestSuite.org.apache.solr.cloud.TestCloudRecovery
Error Message:
Could not find collection:collection1
Stack Trace:
java.lang.AssertionError: Could not find collection:collection1
at __randomizedtesting.SeedInfo.seed([82F7DD912DF45454]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.assertTrue(Assert.java:43)
at org.junit.Assert.assertNotNull(Assert.java:526)
at org.apache.solr.cloud.AbstractDistribZkTestBase.waitForRecoveriesToFinish(AbstractDistribZkTestBase.java:155)
at org.apache.solr.cloud.TestCloudRecovery.setupCluster(TestCloudRecovery.java:70)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:875)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:834)
FAILED: org.apache.solr.cloud.TestTlogReplica.testRecovery
Error Message:
Can not find doc 7 in https://127.0.0.1:61933/solr
Stack Trace:
java.lang.AssertionError: Can not find doc 7 in https://127.0.0.1:61933/solr
at __randomizedtesting.SeedInfo.seed([82F7DD912DF45454:4307A43D00A49EF3]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.assertTrue(Assert.java:43)
at org.junit.Assert.assertNotNull(Assert.java:526)
at org.apache.solr.cloud.TestTlogReplica.checkRTG(TestTlogReplica.java:901)
at org.apache.solr.cloud.TestTlogReplica.testRecovery(TestTlogReplica.java:566)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:834)
Build Log:
[...truncated 1946 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\core\test\temp\junit4-J1-20181006_024440_02814305124661347929342.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 8 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\core\test\temp\junit4-J0-20181006_024440_0296422795196051605808.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 319 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\test-framework\test\temp\junit4-J0-20181006_025641_8367623494765626292136.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\test-framework\test\temp\junit4-J1-20181006_025641_8368185034907871798983.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 1081 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\common\test\temp\junit4-J0-20181006_025741_0608757248334851085621.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\common\test\temp\junit4-J1-20181006_025741_0597993075535058797705.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 257 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\icu\test\temp\junit4-J1-20181006_025942_56416159598526373370992.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\icu\test\temp\junit4-J0-20181006_025942_56412404286683632898801.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 254 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\kuromoji\test\temp\junit4-J1-20181006_025953_60716106615862037426179.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\kuromoji\test\temp\junit4-J0-20181006_025953_6078190905300805034834.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 163 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\morfologik\test\temp\junit4-J0-20181006_030024_12814950341545169451937.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\morfologik\test\temp\junit4-J1-20181006_030024_1284144817945764090867.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 205 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\nori\test\temp\junit4-J1-20181006_030028_8735405803208143994137.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 6 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\nori\test\temp\junit4-J0-20181006_030028_87316841094291457608191.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 172 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\opennlp\test\temp\junit4-J1-20181006_030036_3565404317249700344177.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\opennlp\test\temp\junit4-J0-20181006_030036_3565771848425712707161.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 177 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\phonetic\test\temp\junit4-J1-20181006_030039_95614068483641308927161.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\phonetic\test\temp\junit4-J0-20181006_030039_9564967575081706113794.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 160 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\smartcn\test\temp\junit4-J0-20181006_030049_08410958050921408092250.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\smartcn\test\temp\junit4-J1-20181006_030049_0848848560494683511574.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 168 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\stempel\test\temp\junit4-J1-20181006_030055_54611548584545904896291.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\stempel\test\temp\junit4-J0-20181006_030055_54616933419879906620578.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 204 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\backward-codecs\test\temp\junit4-J0-20181006_030059_68516025740811125005027.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 28 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\backward-codecs\test\temp\junit4-J1-20181006_030059_68511791010390327298443.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 1427 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\benchmark\test\temp\junit4-J1-20181006_030207_5082603567711542731770.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\benchmark\test\temp\junit4-J0-20181006_030207_50810665394896556689476.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 252 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\classification\test\temp\junit4-J0-20181006_030219_14114309278137880322050.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\classification\test\temp\junit4-J1-20181006_030219_14117251330072920566435.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 268 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\codecs\test\temp\junit4-J1-20181006_030235_83617608944076039718761.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\codecs\test\temp\junit4-J0-20181006_030235_836184521474403206735.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 235 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\demo\test\temp\junit4-J0-20181006_030409_43317277903772782990118.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\demo\test\temp\junit4-J1-20181006_030409_4349416108049831778801.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 174 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\expressions\test\temp\junit4-J1-20181006_030414_68114080328012330809846.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\expressions\test\temp\junit4-J0-20181006_030414_6797942786012712932180.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 235 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\facet\test\temp\junit4-J0-20181006_030423_9013402619650758256561.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\facet\test\temp\junit4-J1-20181006_030423_90117506782726354070545.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 183 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\grouping\test\temp\junit4-J0-20181006_030500_44517231346438823113419.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\grouping\test\temp\junit4-J1-20181006_030500_44514323545669743284539.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 249 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\highlighter\test\temp\junit4-J0-20181006_030510_23218031292764038945563.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\highlighter\test\temp\junit4-J1-20181006_030510_2324110590369152082163.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 167 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\join\test\temp\junit4-J0-20181006_030531_73314243500749247902502.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\join\test\temp\junit4-J1-20181006_030531_73410304099512667625421.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 156 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\memory\test\temp\junit4-J0-20181006_030543_88914684575625097398806.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\memory\test\temp\junit4-J1-20181006_030543_8897745216627004486896.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 196 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\misc\test\temp\junit4-J0-20181006_030550_2557853012838686528075.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\misc\test\temp\junit4-J1-20181006_030550_2554253538432394053204.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 339 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\queries\test\temp\junit4-J1-20181006_030612_42816622945585844939202.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\queries\test\temp\junit4-J0-20181006_030612_4282141317208093579300.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 232 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\queryparser\test\temp\junit4-J1-20181006_030622_5928200785696598609835.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\queryparser\test\temp\junit4-J0-20181006_030622_59211326773840009468448.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 217 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\replicator\test\temp\junit4-J1-20181006_030632_09311922093890900358361.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\replicator\test\temp\junit4-J0-20181006_030632_09313943812695216842281.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 214 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\sandbox\test\temp\junit4-J0-20181006_030647_2509712704370486929125.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\sandbox\test\temp\junit4-J1-20181006_030647_25017312228918333195441.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 297 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\spatial-extras\test\temp\junit4-J0-20181006_030726_6773434279930713468885.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\spatial-extras\test\temp\junit4-J1-20181006_030726_67711980492207618362408.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 193 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\spatial3d\test\temp\junit4-J0-20181006_030809_6348512606592091074843.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 5 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\spatial3d\test\temp\junit4-J1-20181006_030809_63413284168478596330503.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 152 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\spatial\test\temp\junit4-J0-20181006_030821_49817396174400647886066.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 256 lines...]
[junit4] JVM J0: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\suggest\test\temp\junit4-J0-20181006_030825_54215060880274744221907.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\suggest\test\temp\junit4-J1-20181006_030825_542408344221670804734.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 2336 lines...]
[junit4] Suite: org.apache.solr.cloud.TestCloudRecovery
[junit4] 2> 1065957 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> Creating dataDir: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\init-core-data-001
[junit4] 2> 1065958 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=false
[junit4] 2> 1065959 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
[junit4] 2> 1065960 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001
[junit4] 2> 1065960 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1065961 INFO (Thread-1755) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1065961 INFO (Thread-1755) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 1065964 ERROR (Thread-1755) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 1066061 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.c.ZkTestServer start zk server on port:57037
[junit4] 2> 1066065 INFO (zkConnectionManagerCallback-2513-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066072 INFO (jetty-launcher-2510-thread-2) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 11+28
[junit4] 2> 1066072 INFO (jetty-launcher-2510-thread-1) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 11+28
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-2) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-2) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-2) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-2) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@155a071c{/solr,null,AVAILABLE}
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-1) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1066073 INFO (jetty-launcher-2510-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3392461{/solr,null,AVAILABLE}
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@78928a9a{SSL,[ssl, http/1.1]}{127.0.0.1:57042}
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.e.j.s.AbstractConnector Started ServerConnector@28deaaaf{SSL,[ssl, http/1.1]}{127.0.0.1:57041}
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.e.j.s.Server Started @1066105ms
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.e.j.s.Server Started @1066105ms
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=57041}
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=57042}
[junit4] 2> 1066077 ERROR (jetty-launcher-2510-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1066077 ERROR (jetty-launcher-2510-thread-2) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 7.6.0
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 7.6.0
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-10-06T03:27:46.929697800Z
[junit4] 2> 1066077 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-10-06T03:27:46.929697800Z
[junit4] 2> 1066080 INFO (zkConnectionManagerCallback-2517-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066080 INFO (zkConnectionManagerCallback-2515-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066081 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1066081 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1066086 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100695555b60002, likely client has closed socket
[junit4] 2> 1066086 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100695555b60001, likely client has closed socket
[junit4] 2> 1066394 INFO (jetty-launcher-2510-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:57037/solr
[junit4] 2> 1066396 INFO (zkConnectionManagerCallback-2521-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066397 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100695555b60003, likely client has closed socket
[junit4] 2> 1066399 INFO (zkConnectionManagerCallback-2523-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066456 INFO (jetty-launcher-2510-thread-2) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:57037/solr
[junit4] 2> 1066458 INFO (zkConnectionManagerCallback-2529-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066459 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100695555b60005, likely client has closed socket
[junit4] 2> 1066461 INFO (zkConnectionManagerCallback-2531-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066472 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:57041_solr
[junit4] 2> 1066472 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.c.Overseer Overseer (id=72173409269055494-127.0.0.1:57041_solr-n_0000000000) starting
[junit4] 2> 1066479 INFO (zkConnectionManagerCallback-2538-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066481 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:57037/solr ready
[junit4] 2> 1066482 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:57041_solr
[junit4] 2> 1066483 INFO (OverseerStateUpdate-72173409269055494-127.0.0.1:57041_solr-n_0000000000) [n:127.0.0.1:57041_solr ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:57041_solr
[junit4] 2> 1066485 INFO (OverseerStateUpdate-72173409269055494-127.0.0.1:57041_solr-n_0000000000) [n:127.0.0.1:57041_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1066487 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1066487 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:57042_solr
[junit4] 2> 1066489 INFO (zkCallback-2537-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1066489 INFO (zkCallback-2522-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1066491 INFO (zkCallback-2530-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1066491 INFO (zkCallback-2537-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1066491 INFO (zkCallback-2522-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1066507 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1066510 INFO (zkConnectionManagerCallback-2544-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066511 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1066511 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:57037/solr ready
[junit4] 2> 1066512 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1066528 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57041.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1066529 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57042.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1066534 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57041.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1066535 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57041.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1066535 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57042.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1066535 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57042.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1066536 INFO (jetty-launcher-2510-thread-2) [n:127.0.0.1:57041_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node2\.
[junit4] 2> 1066536 INFO (jetty-launcher-2510-thread-1) [n:127.0.0.1:57042_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node1\.
[junit4] 2> 1066575 INFO (zkConnectionManagerCallback-2547-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066579 INFO (zkConnectionManagerCallback-2552-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066580 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1066581 INFO (SUITE-TestCloudRecovery-seed#[82F7DD912DF45454]-worker) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:57037/solr ready
[junit4] 2> 1066602 INFO (qtp56340357-9288) [n:127.0.0.1:57041_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params pullReplicas=0&collection.configName=config&maxShardsPerNode=2&name=collection1&nrtReplicas=2&action=CREATE&numShards=2&tlogReplicas=0&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1066605 INFO (OverseerThreadFactory-4472-thread-1-processing-n:127.0.0.1:57041_solr) [n:127.0.0.1:57041_solr ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1
[junit4] 2> 1066714 INFO (OverseerStateUpdate-72173409269055494-127.0.0.1:57041_solr-n_0000000000) [n:127.0.0.1:57041_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"collection1",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"collection1_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:57041/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1066717 INFO (OverseerStateUpdate-72173409269055494-127.0.0.1:57041_solr-n_0000000000) [n:127.0.0.1:57041_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"collection1",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"collection1_shard1_replica_n2",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:57042/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1066719 INFO (OverseerStateUpdate-72173409269055494-127.0.0.1:57041_solr-n_0000000000) [n:127.0.0.1:57041_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"collection1",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"collection1_shard2_replica_n4",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:57041/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1066723 INFO (OverseerStateUpdate-72173409269055494-127.0.0.1:57041_solr-n_0000000000) [n:127.0.0.1:57041_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"collection1",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"collection1_shard2_replica_n6",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:57042/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1066932 ERROR (OverseerThreadFactory-4472-thread-1-processing-n:127.0.0.1:57041_solr) [n:127.0.0.1:57041_solr ] o.a.s.c.a.c.OverseerCollectionMessageHandler Error from shard: https://127.0.0.1:57041/solr
[junit4] 2> org.apache.solr.client.solrj.SolrServerException: IOException occured when talking to server at: https://127.0.0.1:57041/solr
[junit4] 2> at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:657) ~[java/:?]
[junit4] 2> at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255) ~[java/:?]
[junit4] 2> at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244) ~[java/:?]
[junit4] 2> at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219) ~[java/:?]
[junit4] 2> at org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:172) ~[java/:?]
[junit4] 2> at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?]
[junit4] 2> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[?:?]
[junit4] 2> at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?]
[junit4] 2> at com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:176) ~[metrics-core-3.2.6.jar:3.2.6]
[junit4] 2> at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209) [java/:?]
[junit4] 2> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
[junit4] 2> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
[junit4] 2> at java.lang.Thread.run(Thread.java:834) [?:?]
[junit4] 2> Caused by: javax.net.ssl.SSLException: Received fatal alert: internal_error
[junit4] 2> at sun.security.ssl.Alert.createSSLException(Alert.java:129) ~[?:?]
[junit4] 2> at sun.security.ssl.Alert.createSSLException(Alert.java:117) ~[?:?]
[junit4] 2> at sun.security.ssl.TransportContext.fatal(TransportContext.java:308) ~[?:?]
[junit4] 2> at sun.security.ssl.Alert$AlertConsumer.consume(Alert.java:279) ~[?:?]
[junit4] 2> at sun.security.ssl.TransportContext.dispatch(TransportContext.java:181) ~[?:?]
[junit4] 2> at sun.security.ssl.SSLTransport.decode(SSLTransport.java:164) ~[?:?]
[junit4] 2> at sun.security.ssl.SSLSocketImpl.decode(SSLSocketImpl.java:1152) ~[?:?]
[junit4] 2> at sun.security.ssl.SSLSocketImpl.readHandshakeRecord(SSLSocketImpl.java:1063) ~[?:?]
[junit4] 2> at sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:402) ~[?:?]
[junit4] 2> at org.apache.http.conn.ssl.SSLConnectionSocketFactory.createLayeredSocket(SSLConnectionSocketFactory.java:396) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.conn.ssl.SSLConnectionSocketFactory.connectSocket(SSLConnectionSocketFactory.java:355) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.conn.DefaultHttpClientConnectionOperator.connect(DefaultHttpClientConnectionOperator.java:142) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.conn.PoolingHttpClientConnectionManager.connect(PoolingHttpClientConnectionManager.java:359) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.execchain.MainClientExec.establishRoute(MainClientExec.java:381) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:237) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:185) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:89) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:111) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56) ~[httpclient-4.5.3.jar:4.5.3]
[junit4] 2> at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:542) ~[java/:?]
[junit4] 2> ... 12 more
[junit4] 2> 1066941 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=config&newCollection=true&name=collection1_shard1_replica_n1&action=CREATE&numShards=2&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1066941 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 1066950 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=config&newCollection=true&name=collection1_shard1_replica_n2&action=CREATE&numShards=2&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1066956 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=config&newCollection=true&name=collection1_shard2_replica_n6&action=CREATE&numShards=2&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1067954 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1067960 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.s.IndexSchema [collection1_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1067962 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1067962 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_n1' using configuration from collection collection1, trusted=true
[junit4] 2> 1067963 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57041.solr.core.collection1.shard1.replica_n1' (registry 'solr.core.collection1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1067963 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.SolrCore [[collection1_shard1_replica_n1] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node2\collection1_shard1_replica_n1], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node2\.\collection1_shard1_replica_n1\data\]
[junit4] 2> 1067964 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1067967 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1067975 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.s.IndexSchema [collection1_shard1_replica_n2] Schema name=minimal
[junit4] 2> 1067977 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1067977 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_n2' using configuration from collection collection1, trusted=true
[junit4] 2> 1067977 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57042.solr.core.collection1.shard1.replica_n2' (registry 'solr.core.collection1.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1067977 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.c.SolrCore [[collection1_shard1_replica_n2] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node1\collection1_shard1_replica_n2], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node1\.\collection1_shard1_replica_n2\data\]
[junit4] 2> 1067978 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.s.IndexSchema [collection1_shard2_replica_n6] Schema name=minimal
[junit4] 2> 1067980 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1067980 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard2_replica_n6' using configuration from collection collection1, trusted=true
[junit4] 2> 1067981 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57042.solr.core.collection1.shard2.replica_n6' (registry 'solr.core.collection1.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@17459fe0
[junit4] 2> 1067981 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.SolrCore [[collection1_shard2_replica_n6] ] Opening new SolrCore at [C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node1\collection1_shard2_replica_n6], dataDir=[C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\J0\temp\solr.cloud.TestCloudRecovery_82F7DD912DF45454-001\tempDir-001\node1\.\collection1_shard2_replica_n6\data\]
[junit4] 2> 1068258 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1068258 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1068260 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1068260 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1068262 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1068262 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1068265 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1068265 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1068265 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@34584936[collection1_shard2_replica_n6] main]
[junit4] 2> 1068267 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/config
[junit4] 2> 1068268 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/config
[junit4] 2> 1068270 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1068270 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613545046406070272
[junit4] 2> 1068271 INFO (searcherExecutor-4483-thread-1-processing-n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6 c:collection1 s:shard2 r:core_node8) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.SolrCore [collection1_shard2_replica_n6] Registered new searcher Searcher@34584936[collection1_shard2_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1068274 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@1fcddcf[collection1_shard1_replica_n1] main]
[junit4] 2> 1068277 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/config
[junit4] 2> 1068278 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/config
[junit4] 2> 1068278 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={core_node8=0}, version=0}
[junit4] 2> 1068279 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1068279 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613545046415507456
[junit4] 2> 1068280 INFO (searcherExecutor-4481-thread-1-processing-n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1 c:collection1 s:shard1 r:core_node3) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.SolrCore [collection1_shard1_replica_n1] Registered new searcher Searcher@1fcddcf[collection1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1068284 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node3=0}, version=0}
[junit4] 2> 1068284 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard2: total=2 found=1 timeoutin=9999ms
[junit4] 2> 1068291 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard1: total=2 found=1 timeoutin=9999ms
[junit4] 2> 1068318 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1068318 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1068328 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1068328 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1068336 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@4a04d2b9[collection1_shard1_replica_n2] main]
[junit4] 2> 1068341 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/config
[junit4] 2> 1068342 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/config
[junit4] 2> 1068342 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1068342 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1613545046481567744
[junit4] 2> 1068343 INFO (searcherExecutor-4482-thread-1-processing-n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2 c:collection1 s:shard1 r:core_node5) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.c.SolrCore [collection1_shard1_replica_n2] Registered new searcher Searcher@4a04d2b9[collection1_shard1_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1068354 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node3=0, core_node5=0}, version=1}
[junit4] 2> 1068793 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1068793 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1068793 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:57041/solr/collection1_shard1_replica_n1/
[junit4] 2> 1068794 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.u.PeerSync PeerSync: core=collection1_shard1_replica_n1 url=https://127.0.0.1:57041/solr START replicas=[https://127.0.0.1:57042/solr/collection1_shard1_replica_n2/] nUpdates=100
[junit4] 2> 1068794 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.u.PeerSync PeerSync: core=collection1_shard1_replica_n1 url=https://127.0.0.1:57041/solr DONE. We have no versions. sync failed.
[junit4] 2> 1068812 INFO (qtp972162998-9292) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.c.S.Request [collection1_shard1_replica_n2] webapp=/solr path=/get params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 1068813 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the next candidate
[junit4] 2> 1068813 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
[junit4] 2> 1068813 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 1068815 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:57041/solr/collection1_shard1_replica_n1/ shard1
[junit4] 2> 1068970 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1068972 INFO (qtp56340357-9283) [n:127.0.0.1:57041_solr c:collection1 s:shard1 r:core_node3 x:collection1_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node3&collection.configName=config&newCollection=true&name=collection1_shard1_replica_n1&action=CREATE&numShards=2&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2032
[junit4] 2> 1069074 INFO (zkCallback-2530-thread-2) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1069358 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr c:collection1 s:shard1 r:core_node5 x:collection1_shard1_replica_n2] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node5&collection.configName=config&newCollection=true&name=collection1_shard1_replica_n2&action=CREATE&numShards=2&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2408
[junit4] 2> 1069460 INFO (zkCallback-2530-thread-2) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1069460 INFO (zkCallback-2522-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1078301 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext Was waiting for replicas to come up, but they are taking too long - assuming they won't come back till later
[junit4] 2> 1078301 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1078301 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:57042/solr/collection1_shard2_replica_n6/
[junit4] 2> 1078301 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1078301 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.SyncStrategy https://127.0.0.1:57042/solr/collection1_shard2_replica_n6/ has no replicas
[junit4] 2> 1078307 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:57042/solr/collection1_shard2_replica_n6/ shard2
[junit4] 2> 1078308 INFO (zkCallback-2530-thread-2) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1078308 INFO (zkCallback-2522-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1078357 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1078359 INFO (qtp972162998-9296) [n:127.0.0.1:57042_solr c:collection1 s:shard2 r:core_node8 x:collection1_shard2_replica_n6] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node8&collection.configName=config&newCollection=true&name=collection1_shard2_replica_n6&action=CREATE&numShards=2&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=11403
[junit4] 2> 1078360 ERROR (OverseerThreadFactory-4472-thread-1-processing-n:127.0.0.1:57041_solr) [n:127.0.0.1:57041_solr ] o.a.s.c.a.c.OverseerCollectionMessageHandler Cleaning up collection [collection1].
[junit4] 2> 1078360 INFO (OverseerThreadFactory-4472-thread-1-processing-n:127.0.0.1:57041_solr) [n:127.0.0.1:57041_solr ] o.a.s.c.a.c.OverseerCollectionMessageHandler Executing Collection Cmd=action=UNLOAD&deleteInstanceDir=true&deleteDataDir=true&deleteMetricsHistory=true, asyncId=null
[junit4] 2> 1078363 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.collection1.shard1.replica_n2, tag=null
[junit4] 2> 1078363 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@570a3f6c: rootName = solr_57042, domain = solr.core.collection1.shard1.replica_n2, service url = null, agent id = null] for registry solr.core.collection1.shard1.replica_n2 / com.codahale.metrics.MetricRegistry@2e560e19
[junit4] 2> 1078365 INFO (qtp56340357-9286) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.collection1.shard1.replica_n1, tag=null
[junit4] 2> 1078365 INFO (qtp56340357-9286) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@436d43f5: rootName = solr_57041, domain = solr.core.collection1.shard1.replica_n1, service url = null, agent id = null] for registry solr.core.collection1.shard1.replica_n1 / com.codahale.metrics.MetricRegistry@214cc77c
[junit4] 2> 1078374 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.collection1.shard2.replica_n6, tag=null
[junit4] 2> 1078375 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.c.SolrCore [collection1_shard1_replica_n2] CLOSING SolrCore org.apache.solr.core.SolrCore@46a6ea85
[junit4] 2> 1078375 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@14014101: rootName = solr_57042, domain = solr.core.collection1.shard2.replica_n6, service url = null, agent id = null] for registry solr.core.collection1.shard2.replica_n6 / com.codahale.metrics.MetricRegistry@72818ea0
[junit4] 2> 1078376 INFO (qtp56340357-9286) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.c.SolrCore [collection1_shard1_replica_n1] CLOSING SolrCore org.apache.solr.core.SolrCore@434ddb0f
[junit4] 2> 1078376 INFO (qtp56340357-9286) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.collection1.shard1.replica_n1, tag=434ddb0f
[junit4] 2> 1078376 INFO (qtp56340357-9286) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.collection1.shard1.leader, tag=434ddb0f
[junit4] 2> 1078387 INFO (qtp56340357-9286) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node5=0}, version=2}
[junit4] 2> 1078388 INFO (qtp56340357-9286) [n:127.0.0.1:57041_solr x:collection1_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={deleteInstanceDir=true&deleteMetricsHistory=true&core=collection1_shard1_replica_n1&qt=/admin/cores&deleteDataDir=true&action=UNLOAD&wt=javabin&version=2} status=0 QTime=23
[junit4] 2> 1078388 WARN (zkCallback-2522-thread-2) [ ] o.a.s.c.LeaderElector
[junit4] 2> org.apache.solr.common.SolrException: SolrCore not found:collection1_shard1_replica_n2 in []
[junit4] 2> at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:312) ~[java/:?]
[junit4] 2> at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:170) ~[java/:?]
[junit4] 2> at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:135) ~[java/:?]
[junit4] 2> at org.apache.solr.cloud.LeaderElector.access$200(LeaderElector.java:56) ~[java/:?]
[junit4] 2> at org.apache.solr.cloud.LeaderElector$ElectionWatcher.process(LeaderElector.java:348) [java/:?]
[junit4] 2> at org.apache.solr.common.cloud.SolrZkClient$1.lambda$process$1(SolrZkClient.java:270) [java/:?]
[junit4] 2> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) [?:?]
[junit4] 2> at java.util.concurrent.FutureTask.run(FutureTask.java:264) [?:?]
[junit4] 2> at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209) [java/:?]
[junit4] 2> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
[junit4] 2> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
[junit4] 2> at java.lang.Thread.run(Thread.java:834) [?:?]
[junit4] 2> 1078391 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.collection1.shard1.replica_n2, tag=46a6ea85
[junit4] 2> 1078391 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.collection1.shard1.leader, tag=46a6ea85
[junit4] 2> 1078392 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.c.SolrCore [collection1_shard2_replica_n6] CLOSING SolrCore org.apache.solr.core.SolrCore@606ce76e
[junit4] 2> 1078392 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.collection1.shard2.replica_n6, tag=606ce76e
[junit4] 2> 1078392 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.collection1.shard2.leader, tag=606ce76e
[junit4] 2> 1078402 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={}, version=1}
[junit4] 2> 1078403 INFO (qtp972162998-9295) [n:127.0.0.1:57042_solr x:collection1_shard2_replica_n6] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={deleteInstanceDir=true&deleteMetricsHistory=true&core=collection1_shard2_replica_n6&qt=/admin/cores&deleteDataDir=true&action=UNLOAD&wt=javabin&version=2} status=0 QTime=37
[junit4] 2> 1078403 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={}, version=3}
[junit4] 2> 1078403 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContextBase No version found for ephemeral leader parent node, won't remove previous leader registration.
[junit4] 2> 1078404 INFO (qtp972162998-9290) [n:127.0.0.1:57042_solr x:collection1_shard1_replica_n2] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={deleteInstanceDir=true&deleteMetricsHistory=true&core=collection1_shard1_replica_n2&qt=/admin/cores&deleteDataDir=true&action=UNLOAD&wt=javabin&version=2} status=0 QTime=41
[junit4] 2> 1078506 INFO (zkCallback-2522-thread-2) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDeleted path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1078506 INFO (zkCallback-2530-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDeleted path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
[junit4] 2> 1079114 INFO (OverseerThreadFactory-4472-thread-1-processing-n:127.0.0.1:57041_solr) [n:127.0.0.1:57041_solr ] o.a.s.c.a.c.CreateCollectionCmd Cleaned up artifacts for failed create collection for [collection1]
[junit4] 2> 1079116 INFO (qtp56340357-9288) [n:127.0.0.1:57041_solr ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
[junit4] 2> 1080613 INFO (OverseerCollectionConfigSetProcessor-72173409269055494-127.0.0.1:57041_solr-n_0000000000) [n:127.0.0.1:57041_solr ] o.a.s.c.OverseerTas
[...truncated too long message...]
0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-collective-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-event-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-exception-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-journal-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-normalization-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-operational-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-referral-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-schema-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-subtree-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-interceptors-trigger-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-jdbm-partition-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-jdbm1-2.0.0-M2.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-kerberos-codec-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-ldif-partition-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-mavibot-partition-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-protocol-kerberos-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-protocol-ldap-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-protocol-shared-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\apacheds-xdbm-partition-2.0.0-M15.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\api-all-1.0.0-M20.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\bcprov-jdk15on-1.54.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\byte-buddy-1.6.2.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\commons-collections-3.2.2.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\ehcache-core-2.4.4.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\hadoop-common-2.7.4-tests.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\hadoop-hdfs-2.7.4-tests.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\hadoop-minikdc-2.7.4.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\htrace-core-3.2.0-incubating.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\jersey-core-1.9.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\jersey-server-1.9.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\jetty-6.1.26.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\jetty-sslengine-6.1.26.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\jetty-util-6.1.26.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\mina-core-2.0.0-M5.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\mockito-core-2.6.2.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\netty-all-4.0.36.Final.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\core\test-lib\objenesis-2.5.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\analysis\icu\lucene-analyzers-icu-7.6.0-SNAPSHOT.jar;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\contrib\solr-analysis-extras\classes\java;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\queryparser\classes\test;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\build\backward-codecs\classes\test;C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\contrib\analysis-extras\lib\icu4j-62.1.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-launcher.jar;C:\Users\jenkins\.ant\lib\ivy-2.4.0.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-antlr.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-apache-bcel.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-apache-bsf.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-apache-log4j.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-apache-oro.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-apache-regexp.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-apache-resolver.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-apache-xalan2.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-commons-logging.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-commons-net.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-jai.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-javamail.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-jdepend.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-jmf.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-jsch.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-junit.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-junit4.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-netrexx.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-swing.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant-testutil.jar;C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2\lib\ant.jar;C:\Users\jenkins\.ivy2\cache\com.carrotsearch.randomizedtesting\junit4-ant\jars\junit4-ant-2.6.4.jar com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -eventsfile C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\temp\junit4-J0-20181006_031000_74810795491356834586570.events @C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\temp\junit4-J0-20181006_031000_748728328928524463454.suites -stdin
[junit4] ERROR: JVM J0 ended with an exception: Forked process returned with error code: 1. Very likely a JVM crash. See process stdout at: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\temp\junit4-J0-20181006_031000_74813451025012883648907.sysout See process stderr at: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\temp\junit4-J0-20181006_031000_74814315030819690418727.syserr
[junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4.executeSlave(JUnit4.java:1542)
[junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4.access$000(JUnit4.java:123)
[junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:997)
[junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:994)
[junit4] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[junit4] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
[junit4] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
[junit4] at java.base/java.lang.Thread.run(Thread.java:834)
BUILD FAILED
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\build.xml:633: The following error occurred while executing this line:
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\build.xml:577: The following error occurred while executing this line:
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\build.xml:59: The following error occurred while executing this line:
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build.xml:267: The following error occurred while executing this line:
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\common-build.xml:558: The following error occurred while executing this line:
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\common-build.xml:1567: The following error occurred while executing this line:
C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\lucene\common-build.xml:1092: At least one slave process threw an exception, first: Forked process returned with error code: 1. Very likely a JVM crash. See process stdout at: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\temp\junit4-J0-20181006_031000_74813451025012883648907.sysout See process stderr at: C:\Users\jenkins\workspace\Lucene-Solr-7.x-Windows\solr\build\solr-core\test\temp\junit4-J0-20181006_031000_74814315030819690418727.syserr
Total time: 159 minutes 6 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
[WARNINGS] Skipping publisher since build result is FAILURE
Recording test results
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2
Setting ANT_1_8_2_HOME=C:\Users\jenkins\tools\hudson.tasks.Ant_AntInstallation\ANT_1.8.2