Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/151/
Java: 64bit/jdk1.8.0_141 -XX:-UseCompressedOops -XX:+UseParallelGC
1 tests failed.
FAILED: org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.test
Error Message:
Expected 2 of 3 replicas to be active but only found 1;
[core_node5:{"core":"c8n_1x3_lf_shard1_replica_n2","base_url":"http://127.0.0.1:42439/yu/iz","node_name":"127.0.0.1:42439_yu%2Fiz","state":"active","type":"NRT","leader":"true"}];
clusterState: DocCollection(c8n_1x3_lf//clusterstate.json/27)={
"pullReplicas":"0", "replicationFactor":"1", "shards":{"shard1":{
"range":"80000000-7fffffff", "state":"active", "replicas":{
"core_node4":{ "core":"c8n_1x3_lf_shard1_replica_n1",
"base_url":"http://127.0.0.1:43871/yu/iz",
"node_name":"127.0.0.1:43871_yu%2Fiz", "state":"down",
"type":"NRT"}, "core_node5":{
"core":"c8n_1x3_lf_shard1_replica_n2",
"base_url":"http://127.0.0.1:42439/yu/iz",
"node_name":"127.0.0.1:42439_yu%2Fiz", "state":"active",
"type":"NRT", "leader":"true"}, "core_node6":{
"state":"down", "base_url":"http://127.0.0.1:34159/yu/iz",
"core":"c8n_1x3_lf_shard1_replica_n3",
"node_name":"127.0.0.1:34159_yu%2Fiz", "type":"NRT"}}}},
"router":{"name":"compositeId"}, "maxShardsPerNode":"1",
"autoAddReplicas":"false", "nrtReplicas":"3", "tlogReplicas":"0"}
Stack Trace:
java.lang.AssertionError: Expected 2 of 3 replicas to be active but only found
1;
[core_node5:{"core":"c8n_1x3_lf_shard1_replica_n2","base_url":"http://127.0.0.1:42439/yu/iz","node_name":"127.0.0.1:42439_yu%2Fiz","state":"active","type":"NRT","leader":"true"}];
clusterState: DocCollection(c8n_1x3_lf//clusterstate.json/27)={
"pullReplicas":"0",
"replicationFactor":"1",
"shards":{"shard1":{
"range":"80000000-7fffffff",
"state":"active",
"replicas":{
"core_node4":{
"core":"c8n_1x3_lf_shard1_replica_n1",
"base_url":"http://127.0.0.1:43871/yu/iz",
"node_name":"127.0.0.1:43871_yu%2Fiz",
"state":"down",
"type":"NRT"},
"core_node5":{
"core":"c8n_1x3_lf_shard1_replica_n2",
"base_url":"http://127.0.0.1:42439/yu/iz",
"node_name":"127.0.0.1:42439_yu%2Fiz",
"state":"active",
"type":"NRT",
"leader":"true"},
"core_node6":{
"state":"down",
"base_url":"http://127.0.0.1:34159/yu/iz",
"core":"c8n_1x3_lf_shard1_replica_n3",
"node_name":"127.0.0.1:34159_yu%2Fiz",
"type":"NRT"}}}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"1",
"autoAddReplicas":"false",
"nrtReplicas":"3",
"tlogReplicas":"0"}
at
__randomizedtesting.SeedInfo.seed([FE2F13443B09FE7F:767B2C9E95F59387]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.assertTrue(Assert.java:43)
at
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.testRf3WithLeaderFailover(LeaderFailoverAfterPartitionTest.java:169)
at
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.test(LeaderFailoverAfterPartitionTest.java:56)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:985)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:960)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
Build Log:
[...truncated 12624 lines...]
[junit4] Suite: org.apache.solr.cloud.LeaderFailoverAfterPartitionTest
[junit4] 2> 1222112 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[FE2F13443B09FE7F]-worker) [ ]
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks:
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> Creating dataDir:
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/init-core-data-001
[junit4] 2> 1222112 WARN
(SUITE-LeaderFailoverAfterPartitionTest-seed#[FE2F13443B09FE7F]-worker) [ ]
o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=52 numCloses=52
[junit4] 2> 1222113 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[FE2F13443B09FE7F]-worker) [ ]
o.a.s.SolrTestCaseJ4 Using TrieFields (NUMERIC_POINTS_SYSPROP=false)
w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 1222114 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[FE2F13443B09FE7F]-worker) [ ]
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via:
@org.apache.solr.SolrTestCaseJ4$SuppressSSL(bugUrl=https://issues.apache.org/jira/browse/SOLR-5776)
[junit4] 2> 1222114 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[FE2F13443B09FE7F]-worker) [ ]
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /yu/iz
[junit4] 2> 1222115 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1222115 INFO (Thread-2998) [ ] o.a.s.c.ZkTestServer client
port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1222115 INFO (Thread-2998) [ ] o.a.s.c.ZkTestServer
Starting server
[junit4] 2> 1222117 ERROR (Thread-2998) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> 1222215 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.ZkTestServer start zk server on port:45179
[junit4] 2> 1222218 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ]
o.a.z.s.NIOServerCnxn caught end of stream exception
[junit4] 2> EndOfStreamException: Unable to read additional data from
client sessionid 0x15d838d4ed70000, likely client has closed socket
[junit4] 2> at
org.apache.zookeeper.server.NIOServerCnxn.doIO(NIOServerCnxn.java:239)
[junit4] 2> at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:203)
[junit4] 2> at java.lang.Thread.run(Thread.java:748)
[junit4] 2> 1222219 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
to /configs/conf1/solrconfig.xml
[junit4] 2> 1222220 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/schema.xml
to /configs/conf1/schema.xml
[junit4] 2> 1222220 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 1222221 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
to /configs/conf1/stopwords.txt
[junit4] 2> 1222221 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/protwords.txt
to /configs/conf1/protwords.txt
[junit4] 2> 1222221 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/currency.xml
to /configs/conf1/currency.xml
[junit4] 2> 1222222 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
to /configs/conf1/enumsConfig.xml
[junit4] 2> 1222222 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
to /configs/conf1/open-exchange-rates.json
[junit4] 2> 1222222 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 1222223 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
to /configs/conf1/old_synonyms.txt
[junit4] 2> 1222223 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
to /configs/conf1/synonyms.txt
[junit4] 2> 1222224 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly
asked otherwise
[junit4] 2> 1222282 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1222282 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@178fd19{/yu/iz,null,AVAILABLE}
[junit4] 2> 1222283 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@554b0db6{HTTP/1.1,[http/1.1]}{127.0.0.1:39957}
[junit4] 2> 1222283 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server Started @1224148ms
[junit4] 2> 1222283 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/tempDir-001/control/data,
hostContext=/yu/iz, hostPort=40055,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/control-001/cores}
[junit4] 2> 1222283 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1222283 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.1.0
[junit4] 2> 1222283 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1222283 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null, Default config
dir: null
[junit4] 2> 1222283 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-07-27T10:17:42.526Z
[junit4] 2> 1222284 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1222284 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/control-001/solr.xml
[junit4] 2> 1222287 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 1222288 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45179/solr
[junit4] 2> 1222312 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1222312 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.OverseerElectionContext I am going to
be the leader 127.0.0.1:40055_yu%2Fiz
[junit4] 2> 1222312 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.Overseer Overseer
(id=98379409808359428-127.0.0.1:40055_yu%2Fiz-n_0000000000) starting
[junit4] 2> 1222314 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:40055_yu%2Fiz
[junit4] 2> 1222317 INFO
(zkCallback-2329-thread-1-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 1222354 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1222357 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1222357 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1222358 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/control-001/cores
[junit4] 2> 1222373 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1222373 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45179/solr ready
[junit4] 2> 1222374 INFO (SocketProxy-Acceptor-40055) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=37548,localport=40055], receiveBufferSize:531000
[junit4] 2> 1222381 INFO (SocketProxy-Acceptor-40055) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=39957,localport=43758], receiveBufferSize=530904
[junit4] 2> 1222382 INFO (qtp118185642-13094) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
replicationFactor=1&collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:40055_yu%252Fiz&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1222383 INFO
(OverseerThreadFactory-5925-thread-1-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.CreateCollectionCmd Create collection
control_collection
[junit4] 2> 1222485 INFO (SocketProxy-Acceptor-40055) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=37552,localport=40055], receiveBufferSize:531000
[junit4] 2> 1222485 INFO (SocketProxy-Acceptor-40055) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=39957,localport=43762], receiveBufferSize=530904
[junit4] 2> 1222485 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.h.a.CoreAdminOperation core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1222486 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 1222588 INFO
(zkCallback-2329-thread-1-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.c.ZkStateReader A cluster state change:
[WatchedEvent state:SyncConnected type:NodeDataChanged
path:/collections/control_collection/state.json] for collection
[control_collection] has occurred - updating... (live nodes size: [1])
[junit4] 2> 1223501 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.1.0
[junit4] 2> 1223527 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
[junit4] 2> 1223617 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1223623 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1'
using configuration from collection control_collection, trusted=true
[junit4] 2> 1223624 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.control_collection.shard1.replica_n1' (registry
'solr.core.control_collection.shard1.replica_n1') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1223624 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 1223624 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore
at
[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/control-001/cores/control_collection_shard1_replica_n1],
dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/control-001/cores/control_collection_shard1_replica_n1/data/]
[junit4] 2> 1223625 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=36, maxMergeAtOnceExplicit=47, maxMergedSegmentMB=78.650390625,
floorSegmentMB=1.474609375, forceMergeDeletesPctAllowed=3.319400560285881,
segmentsPerTier=48.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
[junit4] 2> 1223627 WARN (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 1223653 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateHandler Using UpdateLog implementation:
org.apache.solr.update.UpdateLog
[junit4] 2> 1223653 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1223654 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1223654 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1223655 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=46, maxMergeAtOnceExplicit=41, maxMergedSegmentMB=19.615234375,
floorSegmentMB=0.859375, forceMergeDeletesPctAllowed=20.104714737153397,
segmentsPerTier=46.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.22900735142783812
[junit4] 2> 1223655 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@38bf0f3e[control_collection_shard1_replica_n1] main]
[junit4] 2> 1223656 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1223656 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1223656 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1223657 INFO
(searcherExecutor-5928-thread-1-processing-n:127.0.0.1:40055_yu%2Fiz
x:control_collection_shard1_replica_n1 s:shard1 c:control_collection)
[n:127.0.0.1:40055_yu%2Fiz c:control_collection s:shard1
x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore
[control_collection_shard1_replica_n1] Registered new searcher
Searcher@38bf0f3e[control_collection_shard1_replica_n1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1223657 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using
new clock 1574070558549606400
[junit4] 2> 1223660 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1223660 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1223661 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy Sync replicas to
http://127.0.0.1:40055/yu/iz/control_collection_shard1_replica_n1/
[junit4] 2> 1223661 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1223661 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy
http://127.0.0.1:40055/yu/iz/control_collection_shard1_replica_n1/ has no
replicas
[junit4] 2> 1223661 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in
election, clear LIR
[junit4] 2> 1223662 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:40055/yu/iz/control_collection_shard1_replica_n1/ shard1
[junit4] 2> 1223763 INFO
(zkCallback-2329-thread-1-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.c.ZkStateReader A cluster state change:
[WatchedEvent state:SyncConnected type:NodeDataChanged
path:/collections/control_collection/state.json] for collection
[control_collection] has occurred - updating... (live nodes size: [1])
[junit4] 2> 1223812 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1223813 INFO (qtp118185642-13096) [n:127.0.0.1:40055_yu%2Fiz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1328
[junit4] 2> 1223814 INFO (qtp118185642-13094) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at
most 30 seconds. Check all shard replicas
[junit4] 2> 1223914 INFO
(zkCallback-2329-thread-1-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.c.ZkStateReader A cluster state change:
[WatchedEvent state:SyncConnected type:NodeDataChanged
path:/collections/control_collection/state.json] for collection
[control_collection] has occurred - updating... (live nodes size: [1])
[junit4] 2> 1224384 INFO
(OverseerCollectionConfigSetProcessor-98379409808359428-127.0.0.1:40055_yu%2Fiz-n_0000000000)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 1224815 INFO (qtp118185642-13094) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={replicationFactor=1&collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:40055_yu%252Fiz&wt=javabin&version=2}
status=0 QTime=2432
[junit4] 2> 1224818 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1224818 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45179/solr ready
[junit4] 2> 1224818 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection
loss:false
[junit4] 2> 1224819 INFO (SocketProxy-Acceptor-40055) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=37566,localport=40055], receiveBufferSize:531000
[junit4] 2> 1224819 INFO (SocketProxy-Acceptor-40055) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=39957,localport=43778], receiveBufferSize=530904
[junit4] 2> 1224820 INFO (qtp118185642-13094) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
replicationFactor=1&collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=1&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1224823 INFO
(OverseerThreadFactory-5925-thread-2-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.CreateCollectionCmd Create collection
collection1
[junit4] 2> 1224825 WARN
(OverseerThreadFactory-5925-thread-2-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.CreateCollectionCmd It is unusual to
create a collection (collection1) without cores.
[junit4] 2> 1225027 INFO (qtp118185642-13094) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at
most 30 seconds. Check all shard replicas
[junit4] 2> 1225027 INFO (qtp118185642-13094) [n:127.0.0.1:40055_yu%2Fiz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={replicationFactor=1&collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=1&wt=javabin&version=2}
status=0 QTime=206
[junit4] 2> 1225075 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-1-001
of type NRT
[junit4] 2> 1225076 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1225076 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@9b25dd2{/yu/iz,null,AVAILABLE}
[junit4] 2> 1225076 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@3c0e99ff{HTTP/1.1,[http/1.1]}{127.0.0.1:45925}
[junit4] 2> 1225077 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server Started @1226942ms
[junit4] 2> 1225077 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/tempDir-001/jetty1,
solrconfig=solrconfig.xml, hostContext=/yu/iz, hostPort=34159,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-1-001/cores}
[junit4] 2> 1225077 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1225077 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.1.0
[junit4] 2> 1225077 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1225077 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null, Default config
dir: null
[junit4] 2> 1225077 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-07-27T10:17:45.320Z
[junit4] 2> 1225078 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1225078 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-1-001/solr.xml
[junit4] 2> 1225080 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 1225082 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45179/solr
[junit4] 2> 1225086 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 1225087 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1225088 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:34159_yu%2Fiz
[junit4] 2> 1225088 INFO (zkCallback-2336-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1225088 INFO
(zkCallback-2329-thread-1-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (1) -> (2)
[junit4] 2> 1225089 INFO
(zkCallback-2341-thread-1-processing-n:127.0.0.1:34159_yu%2Fiz)
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (1) -> (2)
[junit4] 2> 1225153 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1225156 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1225156 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1225157 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-1-001/cores
[junit4] 2> 1225166 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=45326,localport=34159], receiveBufferSize:531000
[junit4] 2> 1225167 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=45925,localport=39970], receiveBufferSize=530904
[junit4] 2> 1225167 INFO (qtp94212645-13142) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with
params
node=127.0.0.1:34159_yu%252Fiz&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1225168 INFO
(OverseerThreadFactory-5925-thread-3-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.AddReplicaCmd Node Identified
127.0.0.1:34159_yu%2Fiz for creating new replica
[junit4] 2> 1225168 INFO
(OverseerCollectionConfigSetProcessor-98379409808359428-127.0.0.1:40055_yu%2Fiz-n_0000000000)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 1225169 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=45330,localport=34159], receiveBufferSize:531000
[junit4] 2> 1225169 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=45925,localport=39974], receiveBufferSize=530904
[junit4] 2> 1225170 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.CoreAdminOperation core create command
qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n1&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1225170 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 1226177 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.SolrConfig
Using Lucene MatchVersion: 7.1.0
[junit4] 2> 1226185 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.s.IndexSchema
[collection1_shard2_replica_n1] Schema name=test
[junit4] 2> 1226286 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.s.IndexSchema
Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1226294 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.CoreContainer
Creating SolrCore 'collection1_shard2_replica_n1' using configuration from
collection collection1, trusted=true
[junit4] 2> 1226294 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.collection1.shard2.replica_n1' (registry
'solr.core.collection1.shard2.replica_n1') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1226294 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.SolrCore
solr.RecoveryStrategy.Builder
[junit4] 2> 1226294 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.SolrCore
[[collection1_shard2_replica_n1] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-1-001/cores/collection1_shard2_replica_n1],
dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-1-001/cores/collection1_shard2_replica_n1/data/]
[junit4] 2> 1226296 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=36, maxMergeAtOnceExplicit=47, maxMergedSegmentMB=78.650390625,
floorSegmentMB=1.474609375, forceMergeDeletesPctAllowed=3.319400560285881,
segmentsPerTier=48.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
[junit4] 2> 1226298 WARN (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 1226327 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1226327 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1226327 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 1226327 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 1226328 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=46, maxMergeAtOnceExplicit=41, maxMergedSegmentMB=19.615234375,
floorSegmentMB=0.859375, forceMergeDeletesPctAllowed=20.104714737153397,
segmentsPerTier=46.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.22900735142783812
[junit4] 2> 1226328 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@6257ee0c[collection1_shard2_replica_n1] main]
[junit4] 2> 1226329 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1226329 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1226329 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1226330 INFO
(searcherExecutor-5939-thread-1-processing-n:127.0.0.1:34159_yu%2Fiz
x:collection1_shard2_replica_n1 s:shard2 c:collection1)
[n:127.0.0.1:34159_yu%2Fiz c:collection1 s:shard2
x:collection1_shard2_replica_n1] o.a.s.c.SolrCore
[collection1_shard2_replica_n1] Registered new searcher
Searcher@6257ee0c[collection1_shard2_replica_n1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1226330 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.u.UpdateLog
Could not find max version in index or recent updates, using new clock
1574070561352450048
[junit4] 2> 1226333 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1226333 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1226333 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.SyncStrategy
Sync replicas to http://127.0.0.1:34159/yu/iz/collection1_shard2_replica_n1/
[junit4] 2> 1226333 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.SyncStrategy
Sync Success - now sync replicas to me
[junit4] 2> 1226333 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.SyncStrategy
http://127.0.0.1:34159/yu/iz/collection1_shard2_replica_n1/ has no replicas
[junit4] 2> 1226333 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in
election, clear LIR
[junit4] 2> 1226334 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1]
o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:34159/yu/iz/collection1_shard2_replica_n1/ shard2
[junit4] 2> 1226484 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.c.ZkController I
am the leader, no recovery necessary
[junit4] 2> 1226485 INFO (qtp94212645-13144) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n1] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n1&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1315
[junit4] 2> 1226490 INFO (qtp94212645-13142) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={node=127.0.0.1:34159_yu%252Fiz&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2}
status=0 QTime=1322
[junit4] 2> 1226548 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-2-001
of type NRT
[junit4] 2> 1226548 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@17a8ac16{/yu/iz,null,AVAILABLE}
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@44605d41{HTTP/1.1,[http/1.1]}{127.0.0.1:33543}
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server Started @1228415ms
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/tempDir-001/jetty2,
solrconfig=solrconfig.xml, hostContext=/yu/iz, hostPort=42439,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-2-001/cores}
[junit4] 2> 1226549 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.1.0
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null, Default config
dir: null
[junit4] 2> 1226549 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-07-27T10:17:46.792Z
[junit4] 2> 1226551 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1226551 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-2-001/solr.xml
[junit4] 2> 1226554 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 1226597 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45179/solr
[junit4] 2> 1226600 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (2)
[junit4] 2> 1226601 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1226601 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:42439_yu%2Fiz
[junit4] 2> 1226602 INFO
(zkCallback-2341-thread-1-processing-n:127.0.0.1:34159_yu%2Fiz)
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (2) -> (3)
[junit4] 2> 1226605 INFO (zkCallback-2336-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1226605 INFO
(zkCallback-2329-thread-2-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (2) -> (3)
[junit4] 2> 1226609 INFO
(zkCallback-2347-thread-1-processing-n:127.0.0.1:42439_yu%2Fiz)
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (2) -> (3)
[junit4] 2> 1226633 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1226636 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1226636 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1226637 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-2-001/cores
[junit4] 2> 1226650 INFO (qtp94212645-13143) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with
params
node=127.0.0.1:42439_yu%252Fiz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1226651 INFO
(OverseerCollectionConfigSetProcessor-98379409808359428-127.0.0.1:40055_yu%2Fiz-n_0000000000)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000004 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 1226651 INFO
(OverseerThreadFactory-5925-thread-4-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.AddReplicaCmd Node Identified
127.0.0.1:42439_yu%2Fiz for creating new replica
[junit4] 2> 1226652 INFO (SocketProxy-Acceptor-42439) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=52072,localport=42439], receiveBufferSize:531000
[junit4] 2> 1226652 INFO (SocketProxy-Acceptor-42439) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=33543,localport=41192], receiveBufferSize=530904
[junit4] 2> 1226653 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
] o.a.s.h.a.CoreAdminOperation core create command
qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n3&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1226654 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 1227663 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.SolrConfig
Using Lucene MatchVersion: 7.1.0
[junit4] 2> 1227676 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.s.IndexSchema
[collection1_shard1_replica_n3] Schema name=test
[junit4] 2> 1227790 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.s.IndexSchema
Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1227798 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.CoreContainer
Creating SolrCore 'collection1_shard1_replica_n3' using configuration from
collection collection1, trusted=true
[junit4] 2> 1227798 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.collection1.shard1.replica_n3' (registry
'solr.core.collection1.shard1.replica_n3') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1227799 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.SolrCore
solr.RecoveryStrategy.Builder
[junit4] 2> 1227799 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.SolrCore
[[collection1_shard1_replica_n3] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-2-001/cores/collection1_shard1_replica_n3],
dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-2-001/cores/collection1_shard1_replica_n3/data/]
[junit4] 2> 1227801 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=36, maxMergeAtOnceExplicit=47, maxMergedSegmentMB=78.650390625,
floorSegmentMB=1.474609375, forceMergeDeletesPctAllowed=3.319400560285881,
segmentsPerTier=48.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
[junit4] 2> 1227803 WARN (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 1227839 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1227839 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1227841 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 1227841 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 1227842 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=46, maxMergeAtOnceExplicit=41, maxMergedSegmentMB=19.615234375,
floorSegmentMB=0.859375, forceMergeDeletesPctAllowed=20.104714737153397,
segmentsPerTier=46.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.22900735142783812
[junit4] 2> 1227842 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@e6a8e77[collection1_shard1_replica_n3] main]
[junit4] 2> 1227842 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1227843 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1227843 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1227844 INFO
(searcherExecutor-5950-thread-1-processing-n:127.0.0.1:42439_yu%2Fiz
x:collection1_shard1_replica_n3 s:shard1 c:collection1)
[n:127.0.0.1:42439_yu%2Fiz c:collection1 s:shard1
x:collection1_shard1_replica_n3] o.a.s.c.SolrCore
[collection1_shard1_replica_n3] Registered new searcher
Searcher@e6a8e77[collection1_shard1_replica_n3]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1227844 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.u.UpdateLog
Could not find max version in index or recent updates, using new clock
1574070562939994112
[junit4] 2> 1227847 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1227847 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1227847 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.SyncStrategy
Sync replicas to http://127.0.0.1:42439/yu/iz/collection1_shard1_replica_n3/
[junit4] 2> 1227847 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.SyncStrategy
Sync Success - now sync replicas to me
[junit4] 2> 1227847 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.SyncStrategy
http://127.0.0.1:42439/yu/iz/collection1_shard1_replica_n3/ has no replicas
[junit4] 2> 1227847 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in
election, clear LIR
[junit4] 2> 1227848 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3]
o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:42439/yu/iz/collection1_shard1_replica_n3/ shard1
[junit4] 2> 1228001 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.c.ZkController I
am the leader, no recovery necessary
[junit4] 2> 1228011 INFO (qtp174884137-13177) [n:127.0.0.1:42439_yu%2Fiz
c:collection1 s:shard1 x:collection1_shard1_replica_n3] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n3&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1358
[junit4] 2> 1228013 INFO (qtp94212645-13143) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={node=127.0.0.1:42439_yu%252Fiz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2}
status=0 QTime=1362
[junit4] 2> 1228088 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 3 in directory
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-3-001
of type NRT
[junit4] 2> 1228089 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1228106 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@622a4e99{/yu/iz,null,AVAILABLE}
[junit4] 2> 1228106 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@56513373{HTTP/1.1,[http/1.1]}{127.0.0.1:43785}
[junit4] 2> 1228106 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.e.j.s.Server Started @1229972ms
[junit4] 2> 1228106 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/tempDir-001/jetty3,
solrconfig=solrconfig.xml, hostContext=/yu/iz, hostPort=43871,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-3-001/cores}
[junit4] 2> 1228106 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1228106 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.1.0
[junit4] 2> 1228106 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1228107 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null, Default config
dir: null
[junit4] 2> 1228107 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-07-27T10:17:48.350Z
[junit4] 2> 1228108 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1228108 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-3-001/solr.xml
[junit4] 2> 1228112 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 1228114 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45179/solr
[junit4] 2> 1228125 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (3)
[junit4] 2> 1228126 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1228127 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:43871_yu%2Fiz
[junit4] 2> 1228127 INFO (zkCallback-2336-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1228127 INFO
(zkCallback-2341-thread-1-processing-n:127.0.0.1:34159_yu%2Fiz)
[n:127.0.0.1:34159_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1228127 INFO
(zkCallback-2347-thread-1-processing-n:127.0.0.1:42439_yu%2Fiz)
[n:127.0.0.1:42439_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1228127 INFO
(zkCallback-2329-thread-2-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1228133 INFO
(zkCallback-2353-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz)
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1228167 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1228169 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1228170 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1228170 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F])
[n:127.0.0.1:43871_yu%2Fiz ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-3-001/cores
[junit4] 2> 1228196 INFO (qtp94212645-13138) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with
params
node=127.0.0.1:43871_yu%252Fiz&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1228197 INFO
(OverseerCollectionConfigSetProcessor-98379409808359428-127.0.0.1:40055_yu%2Fiz-n_0000000000)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000006 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 1228198 INFO
(OverseerThreadFactory-5925-thread-5-processing-n:127.0.0.1:40055_yu%2Fiz)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.AddReplicaCmd Node Identified
127.0.0.1:43871_yu%2Fiz for creating new replica
[junit4] 2> 1228199 INFO (SocketProxy-Acceptor-43871) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=57398,localport=43871], receiveBufferSize:531000
[junit4] 2> 1228199 INFO (SocketProxy-Acceptor-43871) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=43785,localport=60016], receiveBufferSize=530904
[junit4] 2> 1228202 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
] o.a.s.h.a.CoreAdminOperation core create command
qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n5&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1228202 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 1229237 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.c.SolrConfig
Using Lucene MatchVersion: 7.1.0
[junit4] 2> 1229249 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.s.IndexSchema
[collection1_shard2_replica_n5] Schema name=test
[junit4] 2> 1229331 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.s.IndexSchema
Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1229337 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.c.CoreContainer
Creating SolrCore 'collection1_shard2_replica_n5' using configuration from
collection collection1, trusted=true
[junit4] 2> 1229337 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.collection1.shard2.replica_n5' (registry
'solr.core.collection1.shard2.replica_n5') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@663eda4b
[junit4] 2> 1229337 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.c.SolrCore
solr.RecoveryStrategy.Builder
[junit4] 2> 1229337 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.c.SolrCore
[[collection1_shard2_replica_n5] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-3-001/cores/collection1_shard2_replica_n5],
dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_FE2F13443B09FE7F-001/shard-3-001/cores/collection1_shard2_replica_n5/data/]
[junit4] 2> 1229339 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=36, maxMergeAtOnceExplicit=47, maxMergedSegmentMB=78.650390625,
floorSegmentMB=1.474609375, forceMergeDeletesPctAllowed=3.319400560285881,
segmentsPerTier=48.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
[junit4] 2> 1229340 WARN (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 1229366 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1229366 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1229367 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 1229367 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 1229368 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=46, maxMergeAtOnceExplicit=41, maxMergedSegmentMB=19.615234375,
floorSegmentMB=0.859375, forceMergeDeletesPctAllowed=20.104714737153397,
segmentsPerTier=46.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.22900735142783812
[junit4] 2> 1229368 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@5d83b41a[collection1_shard2_replica_n5] main]
[junit4] 2> 1229369 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1229369 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1229369 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1229370 INFO
(searcherExecutor-5961-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2
x:collection1_shard2_replica_n5] o.a.s.c.SolrCore
[collection1_shard2_replica_n5] Registered new searcher
Searcher@5d83b41a[collection1_shard2_replica_n5]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1229370 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.u.UpdateLog
Could not find max version in index or recent updates, using new clock
1574070564540121088
[junit4] 2> 1229372 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.c.ZkController
Core needs to recover:collection1_shard2_replica_n5
[junit4] 2> 1229372 INFO
(updateExecutor-2350-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.DefaultSolrCoreState Running recovery
[junit4] 2> 1229373 INFO (qtp527821299-13209) [n:127.0.0.1:43871_yu%2Fiz
c:collection1 s:shard2 x:collection1_shard2_replica_n5] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n5&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1170
[junit4] 2> 1229383 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.c.RecoveryStrategy Starting recovery
process. recoveringAfterStartup=true
[junit4] 2> 1229383 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.c.RecoveryStrategy ######
startupVersions=[[]]
[junit4] 2> 1229383 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.c.RecoveryStrategy Begin buffering
updates. core=[collection1_shard2_replica_n5]
[junit4] 2> 1229383 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.UpdateLog Starting to buffer updates.
FSUpdateLog{state=ACTIVE, tlog=null}
[junit4] 2> 1229383 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.c.RecoveryStrategy Publishing state of
core [collection1_shard2_replica_n5] as recovering, leader is
[http://127.0.0.1:34159/yu/iz/collection1_shard2_replica_n1/] and I am
[http://127.0.0.1:43871/yu/iz/collection1_shard2_replica_n5/]
[junit4] 2> 1229384 INFO (qtp94212645-13138) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={node=127.0.0.1:43871_yu%252Fiz&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2}
status=0 QTime=1188
[junit4] 2> 1229391 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.SolrTestCaseJ4 ###Starting test
[junit4] 2> 1229391 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase Wait for recoveries to finish - wait
30000 for each attempt
[junit4] 2> 1229391 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[FE2F13443B09FE7F]) [ ]
o.a.s.c.AbstractDistribZkTestBase Wait for recoveries to finish - collection:
collection1 failOnTimeout:true timeout (sec):30000
[junit4] 2> 1229400 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.c.RecoveryStrategy Sending prep recovery
command to [http://127.0.0.1:34159/yu/iz]; [WaitForState:
action=PREPRECOVERY&core=collection1_shard2_replica_n1&nodeName=127.0.0.1:43871_yu%252Fiz&coreNodeName=core_node6&state=recovering&checkLive=true&onlyIfLeader=true&onlyIfLeaderActive=true]
[junit4] 2> 1229401 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=45382,localport=34159], receiveBufferSize:531000
[junit4] 2> 1229404 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=45925,localport=40026], receiveBufferSize=530904
[junit4] 2> 1229406 INFO (qtp94212645-13139) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.PrepRecoveryOp Going to wait for coreNodeName: core_node6, state:
recovering, checkLive: true, onlyIfLeader: true, onlyIfLeaderActive: true,
maxTime: 183 s
[junit4] 2> 1229407 INFO (qtp94212645-13139) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.PrepRecoveryOp In WaitForState(recovering):
collection=collection1, shard=shard2, thisCore=collection1_shard2_replica_n1,
leaderDoesNotNeedRecovery=false, isLeader? true, live=true, checkLive=true,
currentState=down, localState=active, nodeName=127.0.0.1:43871_yu%2Fiz,
coreNodeName=core_node6, onlyIfActiveCheckResult=false, nodeProps:
core_node6:{"core":"collection1_shard2_replica_n5","base_url":"http://127.0.0.1:43871/yu/iz","node_name":"127.0.0.1:43871_yu%2Fiz","state":"down","type":"NRT"}
[junit4] 2> 1230198 INFO
(OverseerCollectionConfigSetProcessor-98379409808359428-127.0.0.1:40055_yu%2Fiz-n_0000000000)
[n:127.0.0.1:40055_yu%2Fiz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000008 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 1230407 INFO (qtp94212645-13139) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.PrepRecoveryOp In WaitForState(recovering):
collection=collection1, shard=shard2, thisCore=collection1_shard2_replica_n1,
leaderDoesNotNeedRecovery=false, isLeader? true, live=true, checkLive=true,
currentState=recovering, localState=active, nodeName=127.0.0.1:43871_yu%2Fiz,
coreNodeName=core_node6, onlyIfActiveCheckResult=false, nodeProps:
core_node6:{"core":"collection1_shard2_replica_n5","base_url":"http://127.0.0.1:43871/yu/iz","node_name":"127.0.0.1:43871_yu%2Fiz","state":"recovering","type":"NRT"}
[junit4] 2> 1230407 INFO (qtp94212645-13139) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.h.a.PrepRecoveryOp Waited coreNodeName: core_node6, state: recovering,
checkLive: true, onlyIfLeader: true for: 1 seconds.
[junit4] 2> 1230407 INFO (qtp94212645-13139) [n:127.0.0.1:34159_yu%2Fiz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores
params={nodeName=127.0.0.1:43871_yu%252Fiz&onlyIfLeaderActive=true&core=collection1_shard2_replica_n1&coreNodeName=core_node6&action=PREPRECOVERY&checkLive=true&state=recovering&onlyIfLeader=true&wt=javabin&version=2}
status=0 QTime=1000
[junit4] 2> 1230907 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.c.RecoveryStrategy Attempting to
PeerSync from [http://127.0.0.1:34159/yu/iz/collection1_shard2_replica_n1/] -
recoveringAfterStartup=[true]
[junit4] 2> 1230908 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.PeerSync PeerSync:
core=collection1_shard2_replica_n5 url=http://127.0.0.1:43871/yu/iz START
replicas=[http://127.0.0.1:34159/yu/iz/collection1_shard2_replica_n1/]
nUpdates=100
[junit4] 2> 1230909 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=45396,localport=34159], receiveBufferSize:531000
[junit4] 2> 1230913 INFO (SocketProxy-Acceptor-34159) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=45925,localport=40040], receiveBufferSize=530904
[junit4] 2> 1230914 INFO (qtp94212645-13140) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 r:core_node2 x:collection1_shard2_replica_n1]
o.a.s.u.IndexFingerprint IndexFingerprint millis:0.0
result:{maxVersionSpecified=9223372036854775807, maxVersionEncountered=0,
maxInHash=0, versionsHash=0, numVersions=0, numDocs=0, maxDoc=0}
[junit4] 2> 1230914 INFO (qtp94212645-13140) [n:127.0.0.1:34159_yu%2Fiz
c:collection1 s:shard2 r:core_node2 x:collection1_shard2_replica_n1]
o.a.s.c.S.Request [collection1_shard2_replica_n1] webapp=/yu/iz path=/get
params={distrib=false&qt=/get&getFingerprint=9223372036854775807&wt=javabin&version=2}
status=0 QTime=0
[junit4] 2> 1230914 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.IndexFingerprint IndexFingerprint
millis:0.0 result:{maxVersionSpecified=9223372036854775807,
maxVersionEncountered=0, maxInHash=0, versionsHash=0, numVersions=0, numDocs=0,
maxDoc=0}
[junit4] 2> 1230914 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.PeerSync We are already in sync. No
need to do a PeerSync
[junit4] 2> 1230914 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.DirectUpdateHandler2 start
commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1230914 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.DirectUpdateHandler2 No uncommitted
changes. Skipping IW.commit.
[junit4] 2> 1230914 INFO
(recoveryExecutor-2351-thread-1-processing-n:127.0.0.1:43871_yu%2Fiz
x:collection1_shard2_replica_n5 s:shard2 c:collection1 r:core_node6)
[n:127.0.0.1:43871_yu%2Fiz c:collection1 s:shard2 r:core_node6
x:collection1_shard2_replica_n5] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1230914 INFO (re
[...truncated too long message...]
'parallelStream' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 8. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/highlight/DefaultSolrHighlighter.java
(at line 578)
[ecj-lint] tvWindowStream = new OffsetWindowTokenFilter(tvStream);
[ecj-lint] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: 'tvWindowStream' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 9. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/request/SimpleFacets.java
(at line 912)
[ecj-lint] fastForRandomSet = new HashDocSet(sset.getDocs(), 0,
sset.size());
[ecj-lint]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: 'fastForRandomSet' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 10. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/response/SmileResponseWriter.java
(at line 33)
[ecj-lint] new SmileWriter(out, request, response).writeResponse();
[ecj-lint] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: '<unassigned Closeable value>' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 11. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/schema/OpenExchangeRatesOrgProvider.java
(at line 146)
[ecj-lint] ratesJsonStream =
resourceLoader.openResource(ratesFileLocation);
[ecj-lint] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: 'ratesJsonStream' is not closed at this location
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 12. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/search/CollapsingQParserPlugin.java
(at line 1315)
[ecj-lint] UninvertingReader uninvertingReader = new UninvertingReader(new
ReaderWrapper(searcher.getSlowAtomicReader(), collapseField), mapping);
[ecj-lint] ^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: 'uninvertingReader' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 13. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/search/DocSlice.java
(at line 145)
[ecj-lint] HashDocSet h = new HashDocSet(docs,offset,len);
[ecj-lint] ^
[ecj-lint] Resource leak: 'h' is never closed
[ecj-lint] ----------
[ecj-lint] 14. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/search/DocSlice.java
(at line 154)
[ecj-lint] HashDocSet h = new HashDocSet(docs,offset,len);
[ecj-lint] ^
[ecj-lint] Resource leak: 'h' is never closed
[ecj-lint] ----------
[ecj-lint] 15. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/search/DocSlice.java
(at line 163)
[ecj-lint] HashDocSet h = new HashDocSet(docs,offset,len);
[ecj-lint] ^
[ecj-lint] Resource leak: 'h' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 16. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/search/function/OrdFieldSource.java
(at line 78)
[ecj-lint] SolrIndexSearcher is = (SolrIndexSearcher) o;
[ecj-lint] ^^
[ecj-lint] Resource leak: 'is' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 17. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/search/function/ReverseOrdFieldSource.java
(at line 78)
[ecj-lint] SolrIndexSearcher is = (SolrIndexSearcher) o;
[ecj-lint] ^^
[ecj-lint] Resource leak: 'is' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 18. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/servlet/SolrDispatchFilter.java
(at line 287)
[ecj-lint] throw new SolrException(ErrorCode.SERVER_ERROR, "Error occurred
while loading solr.xml from zookeeper", e);
[ecj-lint]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: 'loader' is not closed at this location
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 19. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/CdcrTransactionLog.java
(at line 98)
[ecj-lint] ChannelFastInputStream is = new ChannelFastInputStream(channel,
pos);
[ecj-lint] ^^
[ecj-lint] Resource leak: 'is' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 20. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/HdfsTransactionLog.java
(at line 191)
[ecj-lint] LogCodec codec = new LogCodec(resolver);
[ecj-lint] ^^^^^
[ecj-lint] Resource leak: 'codec' is never closed
[ecj-lint] ----------
[ecj-lint] 21. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/HdfsTransactionLog.java
(at line 261)
[ecj-lint] LogCodec codec = new LogCodec(resolver);
[ecj-lint] ^^^^^
[ecj-lint] Resource leak: 'codec' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 22. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/TransactionLog.java
(at line 228)
[ecj-lint] ChannelFastInputStream is = new ChannelFastInputStream(channel,
pos);
[ecj-lint] ^^
[ecj-lint] Resource leak: 'is' is never closed
[ecj-lint] ----------
[ecj-lint] 23. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/TransactionLog.java
(at line 228)
[ecj-lint] ChannelFastInputStream is = new ChannelFastInputStream(channel,
pos);
[ecj-lint] ^^
[ecj-lint] Resource leak: 'is' is never closed
[ecj-lint] ----------
[ecj-lint] 24. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/TransactionLog.java
(at line 260)
[ecj-lint] LogCodec codec = new LogCodec(resolver);
[ecj-lint] ^^^^^
[ecj-lint] Resource leak: 'codec' is never closed
[ecj-lint] ----------
[ecj-lint] 25. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/TransactionLog.java
(at line 275)
[ecj-lint] LogCodec codec = new LogCodec(resolver);
[ecj-lint] ^^^^^
[ecj-lint] Resource leak: 'codec' is never closed
[ecj-lint] ----------
[ecj-lint] 26. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/TransactionLog.java
(at line 529)
[ecj-lint] LogCodec codec = new LogCodec(resolver);
[ecj-lint] ^^^^^
[ecj-lint] Resource leak: 'codec' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 27. ERROR in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/processor/UUIDUpdateProcessorFactory.java
(at line 23)
[ecj-lint] import org.apache.solr.common.SolrException;
[ecj-lint] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] The import org.apache.solr.common.SolrException is never used
[ecj-lint] ----------
[ecj-lint] 28. ERROR in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/processor/UUIDUpdateProcessorFactory.java
(at line 24)
[ecj-lint] import static org.apache.solr.common.SolrException.ErrorCode.*;
[ecj-lint] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] The import org.apache.solr.common.SolrException.ErrorCode is never
used
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 29. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/update/processor/UpdateRequestProcessor.java
(at line 85)
[ecj-lint] UpdateRequestProcessor p = this;
[ecj-lint] ^
[ecj-lint] Resource leak: 'p' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 30. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/util/SimplePostTool.java
(at line 655)
[ecj-lint] BAOS bos = new BAOS();
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'bos' is never closed
[ecj-lint] ----------
[ecj-lint] 31. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/util/SimplePostTool.java
(at line 655)
[ecj-lint] BAOS bos = new BAOS();
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'bos' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 32. WARNING in
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/java/org/apache/solr/util/SolrLogLayout.java
(at line 149)
[ecj-lint] SolrQueryRequest req = requestInfo == null ? null :
requestInfo.getReq();
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'req' is never closed
[ecj-lint] ----------
[ecj-lint] 32 problems (2 errors, 30 warnings)
BUILD FAILED
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/build.xml:810: The following
error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/build.xml:101: The following
error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build.xml:689: The following
error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/lucene/common-build.xml:2007: The
following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/lucene/common-build.xml:2046:
Compile failed; see the compiler error output for details.
Total time: 63 minutes 16 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
[WARNINGS] Skipping publisher since build result is FAILURE
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]