Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/389/
3 tests failed.
FAILED: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest.test
Error Message:
There are still nodes recoverying - waited for 320 seconds
Stack Trace:
java.lang.AssertionError: There are still nodes recoverying - waited for 320
seconds
at
__randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F:ECB058A35192A2A7]:0)
at org.junit.Assert.fail(Assert.java:93)
at
org.apache.solr.cloud.AbstractDistribZkTestBase.waitForRecoveriesToFinish(AbstractDistribZkTestBase.java:185)
at
org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForRecoveriesToFinish(AbstractFullDistribZkTestBase.java:920)
at
org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForThingsToLevelOut(AbstractFullDistribZkTestBase.java:1477)
at
org.apache.solr.cloud.RestartWhileUpdatingTest.test(RestartWhileUpdatingTest.java:145)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1010)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:985)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
FAILED:
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest
Error Message:
7 threads leaked from SUITE scope at
org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest: 1)
Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 2) Thread[id=40196,
name=searcherExecutor-12526-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 3) Thread[id=40237,
name=searcherExecutor-12540-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 4) Thread[id=40318,
name=searcherExecutor-12568-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 5) Thread[id=40399,
name=searcherExecutor-12588-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 6) Thread[id=40274,
name=searcherExecutor-12554-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 7) Thread[id=40434,
name=searcherExecutor-12602-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 7 threads leaked from SUITE
scope at org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest:
1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
at __randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F]:0)
FAILED:
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest
Error Message:
There are still zombie threads that couldn't be terminated: 1)
Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 2) Thread[id=40196,
name=searcherExecutor-12526-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 3) Thread[id=40237,
name=searcherExecutor-12540-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 4) Thread[id=40318,
name=searcherExecutor-12568-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 5) Thread[id=40399,
name=searcherExecutor-12588-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 6) Thread[id=40274,
name=searcherExecutor-12554-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 7) Thread[id=40434,
name=searcherExecutor-12602-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native
Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie
threads that couldn't be terminated:
1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING,
group=TGRP-HdfsRestartWhileUpdatingTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
at __randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F]:0)
Build Log:
[...truncated 15263 lines...]
[junit4] Suite: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest
[junit4] 2> Creating dataDir:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/init-core-data-001
[junit4] 2> 3455430 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true)
w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 3455433 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (true) via:
@org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
[junit4] 2> 3455433 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks:
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> 3455433 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /pq_/z
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 3455475 WARN
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 3455481 WARN
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 3455482 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log jetty-6.1.26
[junit4] 2> 3455502 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log Extract
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs
to ./temp/Jetty_localhost_38023_hdfs____.tovdyi/webapp
[junit4] 2> 3455926 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log Started
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:38023
[junit4] 2> 3456003 WARN
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 3456005 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log jetty-6.1.26
[junit4] 2> 3456025 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log Extract
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
to ./temp/Jetty_localhost_36563_datanode____1av2u/webapp
[junit4] 2> 3456424 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log Started
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:36563
[junit4] 2> 3456483 WARN
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 3456484 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log jetty-6.1.26
[junit4] 2> 3456511 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log Extract
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
to ./temp/Jetty_localhost_41352_datanode____rzdyvu/webapp
[junit4] 2> 3456568 ERROR (DataNode:
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data1/,
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data2/]]
heartbeating to localhost/127.0.0.1:34159) [ ]
o.a.h.h.s.d.DirectoryScanner
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1
ms/sec. Assuming default value of 1000
[junit4] 2> 3456580 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xfe700a2551df0: from storage
DS-8dc247b8-210c-4bb7-81f7-056a390b655d node
DatanodeRegistration(127.0.0.1:35883,
datanodeUuid=21573dcb-4239-46df-97c5-88f1d59ce34d, infoPort=35438,
infoSecurePort=0, ipcPort=33352,
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0,
hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 3456580 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xfe700a2551df0: from storage
DS-3c8bf2dd-d4d4-4ba3-89ef-97f75d414464 node
DatanodeRegistration(127.0.0.1:35883,
datanodeUuid=21573dcb-4239-46df-97c5-88f1d59ce34d, infoPort=35438,
infoSecurePort=0, ipcPort=33352,
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0,
hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 3457073 INFO
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ]
o.m.log Started
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41352
[junit4] 2> 3457173 ERROR (DataNode:
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data3/,
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data4/]]
heartbeating to localhost/127.0.0.1:34159) [ ]
o.a.h.h.s.d.DirectoryScanner
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1
ms/sec. Assuming default value of 1000
[junit4] 2> 3457185 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xfe700c65df150: from storage
DS-12755719-cd60-4944-830c-54f7cb39a54f node
DatanodeRegistration(127.0.0.1:46625,
datanodeUuid=f2212e0f-6033-4d4b-8cc7-b19f374e3835, infoPort=43148,
infoSecurePort=0, ipcPort=36074,
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0,
hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 3457185 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xfe700c65df150: from storage
DS-d3aaa494-7d27-41ae-a4b5-5906476e046f node
DatanodeRegistration(127.0.0.1:46625,
datanodeUuid=f2212e0f-6033-4d4b-8cc7-b19f374e3835, infoPort=43148,
infoSecurePort=0, ipcPort=36074,
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0,
hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 3457273 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 3457274 INFO (Thread-12411) [ ] o.a.s.c.ZkTestServer
client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 3457274 INFO (Thread-12411) [ ] o.a.s.c.ZkTestServer
Starting server
[junit4] 2> 3457276 ERROR (Thread-12411) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> 3457374 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ZkTestServer start zk server on port:44631
[junit4] 2> 3457377 INFO (zkConnectionManagerCallback-10510-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3457382 INFO (zkConnectionManagerCallback-10512-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3457388 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
to /configs/conf1/solrconfig.xml
[junit4] 2> 3457391 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/schema15.xml
to /configs/conf1/schema.xml
[junit4] 2> 3457392 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 3457394 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
to /configs/conf1/stopwords.txt
[junit4] 2> 3457396 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/protwords.txt
to /configs/conf1/protwords.txt
[junit4] 2> 3457397 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/currency.xml
to /configs/conf1/currency.xml
[junit4] 2> 3457399 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
to /configs/conf1/enumsConfig.xml
[junit4] 2> 3457400 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
to /configs/conf1/open-exchange-rates.json
[junit4] 2> 3457402 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 3457403 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
to /configs/conf1/old_synonyms.txt
[junit4] 2> 3457405 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
to /configs/conf1/synonyms.txt
[junit4] 2> 3457407 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly
asked otherwise
[junit4] 2> 3457590 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git:
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
[junit4] 2> 3457591 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 3457591 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 3457591 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 3457594 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@2a9d160f{/pq_/z,null,AVAILABLE}
[junit4] 2> 3457594 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@3f91f18a{HTTP/1.1,[http/1.1]}{127.0.0.1:36044}
[junit4] 2> 3457594 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.Server Started @3457651ms
[junit4] 2> 3457594 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_control_data,
replicaType=NRT, hostContext=/pq_/z, hostPort=36044,
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores}
[junit4] 2> 3457594 ERROR
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 3457594 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter Using logger factory
org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 3457595 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.7.0
[junit4] 2> 3457595 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 3457595 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 3457595 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2018-11-27T17:55:33.178Z
[junit4] 2> 3457597 INFO (zkConnectionManagerCallback-10514-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3457599 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 3457599 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/solr.xml
[junit4] 2> 3457603 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay
is ignored
[junit4] 2> 3457603 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter
autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 3457605 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 3457724 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr
[junit4] 2> 3457726 INFO (zkConnectionManagerCallback-10518-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3457731 INFO (zkConnectionManagerCallback-10520-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3457867 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerElectionContext I am going to
be the leader 127.0.0.1:36044_pq_%2Fz
[junit4] 2> 3457868 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.Overseer Overseer
(id=72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) starting
[junit4] 2> 3457876 INFO (zkConnectionManagerCallback-10527-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3457878 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.s.i.ZkClientClusterStateProvider
Cluster at 127.0.0.1:44631/solr ready
[junit4] 2> 3457879 INFO
(OverseerStateUpdate-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.Overseer Starting to work on the main
queue : 127.0.0.1:36044_pq_%2Fz
[junit4] 2> 3457879 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:36044_pq_%2Fz
[junit4] 2> 3457888 INFO
(OverseerStateUpdate-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 3457891 INFO (zkCallback-10526-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 3457896 DEBUG
(OverseerAutoScalingTriggerThread-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
[ ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial
livenodes: [127.0.0.1:36044_pq_%2Fz]
[junit4] 2> 3457899 DEBUG (ScheduledTrigger-12519-thread-2) [ ]
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with
currently live nodes: 1
[junit4] 2> 3457911 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.h.a.MetricsHistoryHandler No .system
collection, keeping metrics history in memory.
[junit4] 2> 3457941 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3457958 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3457958 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3457960 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F])
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions underneath
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores
[junit4] 2> 3457998 INFO (zkConnectionManagerCallback-10533-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3457999 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 3458001 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
[junit4] 2> 3458002 INFO (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:36044_pq_%252Fz&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 3458005 INFO
(OverseerThreadFactory-12521-thread-1-processing-n:127.0.0.1:36044_pq_%2Fz)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.a.c.CreateCollectionCmd Create
collection control_collection
[junit4] 2> 3458112 INFO (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
status=0 QTime=0
[junit4] 2> 3458116 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core
create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 3458117 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault
Allocating transient cache for 4 transient cores
[junit4] 2> 3458902 DEBUG (ScheduledTrigger-12519-thread-3) [ ]
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with
currently live nodes: 1
[junit4] 2> 3459143 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.7.0
[junit4] 2> 3459162 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
[junit4] 2> 3459268 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 3459289 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1'
using configuration from collection control_collection, trusted=true
[junit4] 2> 3459290 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.control_collection.shard1.replica_n1' (registry
'solr.core.control_collection.shard1.replica_n1') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3459290 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore
at
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores/control_collection_shard1_replica_n1],
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores/control_collection_shard1_replica_n1/data/]
[junit4] 2> 3459295 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=47, maxMergeAtOnceExplicit=22, maxMergedSegmentMB=1.4248046875,
floorSegmentMB=0.9541015625, forceMergeDeletesPctAllowed=16.83237086906014,
segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.4209624511310861, deletesPctAllowed=39.50145037974311
[junit4] 2> 3459304 WARN (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 3459383 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateHandler Using UpdateLog implementation:
org.apache.solr.update.UpdateLog
[junit4] 2> 3459383 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 3459385 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.CommitTracker Hard AutoCommit: if uncommitted for 30000ms;
[junit4] 2> 3459385 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.CommitTracker Soft AutoCommit: if uncommitted for 3000ms;
[junit4] 2> 3459387 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy:
minMergeSize=1677721, mergeFactor=41, maxMergeSize=2147483648,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.0]
[junit4] 2> 3459388 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@1dd69426[control_collection_shard1_replica_n1] main]
[junit4] 2> 3459389 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 3459390 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 3459391 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 3459391 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using
new clock 1618310683661697024
[junit4] 2> 3459393 INFO
(searcherExecutor-12526-thread-1-processing-n:127.0.0.1:36044_pq_%2Fz
x:control_collection_shard1_replica_n1 c:control_collection s:shard1)
[n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1
x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore
[control_collection_shard1_replica_n1] Registered new searcher
Searcher@1dd69426[control_collection_shard1_replica_n1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 3459401 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ZkShardTerms Successful update of terms at
/collections/control_collection/terms/shard1 to Terms{values={core_node2=0},
version=0}
[junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy Sync replicas to
http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/
[junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy
http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/ has no
replicas
[junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in
election, clear LIR
[junit4] 2> 3459412 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/ shard1
[junit4] 2> 3459515 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 3459523 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1406
[junit4] 2> 3459526 INFO (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at
most 30 seconds. Check all shard replicas
[junit4] 2> 3459618 INFO (zkCallback-10519-thread-1) [ ]
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent
state:SyncConnected type:NodeDataChanged
path:/collections/control_collection/state.json] for collection
[control_collection] has occurred - updating... (live nodes size: [1])
[junit4] 2> 3459902 DEBUG (ScheduledTrigger-12519-thread-4) [ ]
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with
currently live nodes: 1
[junit4] 2> 3460008 INFO
(OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 3460527 INFO (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:36044_pq_%252Fz&wt=javabin&version=2}
status=0 QTime=2524
[junit4] 2> 3460532 INFO (zkConnectionManagerCallback-10538-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3460534 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 3460535 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
[junit4] 2> 3460535 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection
loss:false
[junit4] 2> 3460536 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 3460539 INFO
(OverseerThreadFactory-12521-thread-2-processing-n:127.0.0.1:36044_pq_%2Fz)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.a.c.CreateCollectionCmd Create
collection collection1
[junit4] 2> 3460744 WARN
(OverseerThreadFactory-12521-thread-2-processing-n:127.0.0.1:36044_pq_%2Fz)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.a.c.CreateCollectionCmd It is unusual
to create a collection (collection1) without cores.
[junit4] 2> 3460747 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at
most 30 seconds. Check all shard replicas
[junit4] 2> 3460748 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2}
status=0 QTime=211
[junit4] 2> 3460902 DEBUG (ScheduledTrigger-12519-thread-3) [ ]
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with
currently live nodes: 1
[junit4] 2> 3460927 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001
of type NRT
[junit4] 2> 3460931 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git:
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
[junit4] 2> 3460933 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 3460933 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 3460934 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 3460934 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@4682a57{/pq_/z,null,AVAILABLE}
[junit4] 2> 3460934 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@325d81d3{HTTP/1.1,[http/1.1]}{127.0.0.1:44392}
[junit4] 2> 3460934 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.Server Started @3460991ms
[junit4] 2> 3460934 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_jetty1,
replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/pq_/z,
hostPort=44392,
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores}
[junit4] 2> 3460935 ERROR
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 3460937 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter Using logger factory
org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 3460937 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.7.0
[junit4] 2> 3460937 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 3460937 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 3460937 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2018-11-27T17:55:36.520Z
[junit4] 2> 3460939 INFO (zkConnectionManagerCallback-10540-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3460940 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 3460940 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/solr.xml
[junit4] 2> 3460945 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay
is ignored
[junit4] 2> 3460946 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter
autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 3460948 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 3461155 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr
[junit4] 2> 3461157 INFO (zkConnectionManagerCallback-10544-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3461159 INFO (zkConnectionManagerCallback-10546-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3461168 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 3461173 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 3461174 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:44392_pq_%2Fz
[junit4] 2> 3461175 INFO (zkCallback-10526-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 3461182 INFO (zkCallback-10519-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 3461182 INFO (zkCallback-10537-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 3461182 INFO (zkCallback-10545-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 3461204 INFO (zkConnectionManagerCallback-10553-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3461205 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 3461206 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
[junit4] 2> 3461206 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history
in memory.
[junit4] 2> 3461237 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node')
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3461255 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm')
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3461255 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry
'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3461258 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores
[junit4] 2> 3461299 INFO (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with
params
node=127.0.0.1:44392_pq_%252Fz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 3461303 INFO
(OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 3461313 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics
params={wt=javabin&version=2&key=solr.core.control_collection.shard1.replica_n1:INDEX.sizeInBytes}
status=0 QTime=2
[junit4] 2> 3461315 INFO (qtp127098813-40157) [n:127.0.0.1:36044_pq_%2Fz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
status=0 QTime=1
[junit4] 2> 3461318 INFO (qtp641447673-40209) [n:127.0.0.1:44392_pq_%2Fz
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
status=0 QTime=0
[junit4] 2> 3461319 INFO
(OverseerThreadFactory-12521-thread-3-processing-n:127.0.0.1:36044_pq_%2Fz)
[n:127.0.0.1:36044_pq_%2Fz c:collection1 s:shard1 ] o.a.s.c.a.c.AddReplicaCmd
Node Identified 127.0.0.1:44392_pq_%2Fz for creating new replica of shard
shard1 for collection collection1
[junit4] 2> 3461324 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
x:collection1_shard1_replica_n21] o.a.s.h.a.CoreAdminOperation core create
command
qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 3461903 DEBUG (ScheduledTrigger-12519-thread-4) [ ]
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with
currently live nodes: 2
[junit4] 2> 3462345 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrConfig
Using Lucene MatchVersion: 7.7.0
[junit4] 2> 3462361 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema
[collection1_shard1_replica_n21] Schema name=test
[junit4] 2> 3462456 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema
Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 3462474 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.CoreContainer
Creating SolrCore 'collection1_shard1_replica_n21' using configuration from
collection collection1, trusted=true
[junit4] 2> 3462475 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.collection1.shard1.replica_n21' (registry
'solr.core.collection1.shard1.replica_n21') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3462475 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrCore
[[collection1_shard1_replica_n21] ] Opening new SolrCore at
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores/collection1_shard1_replica_n21],
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores/collection1_shard1_replica_n21/data/]
[junit4] 2> 3462480 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=47, maxMergeAtOnceExplicit=22, maxMergedSegmentMB=1.4248046875,
floorSegmentMB=0.9541015625, forceMergeDeletesPctAllowed=16.83237086906014,
segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.4209624511310861, deletesPctAllowed=39.50145037974311
[junit4] 2> 3462497 WARN (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 3462568 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 3462569 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 3462570 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker
Hard AutoCommit: if uncommitted for 30000ms;
[junit4] 2> 3462570 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker
Soft AutoCommit: if uncommitted for 3000ms;
[junit4] 2> 3462573 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy:
minMergeSize=1677721, mergeFactor=41, maxMergeSize=2147483648,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.0]
[junit4] 2> 3462573 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@5fc5a521[collection1_shard1_replica_n21] main]
[junit4] 2> 3462575 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 3462575 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 3462576 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 3462577 INFO
(searcherExecutor-12540-thread-1-processing-n:127.0.0.1:44392_pq_%2Fz
x:collection1_shard1_replica_n21 c:collection1 s:shard1)
[n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1
x:collection1_shard1_replica_n21] o.a.s.c.SolrCore
[collection1_shard1_replica_n21] Registered new searcher
Searcher@5fc5a521[collection1_shard1_replica_n21]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 3462577 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog
Could not find max version in index or recent updates, using new clock
1618310687002460160
[junit4] 2> 3462584 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ZkShardTerms
Successful update of terms at /collections/collection1/terms/shard1 to
Terms{values={core_node22=0}, version=0}
[junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy
Sync replicas to http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/
[junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy
Sync Success - now sync replicas to me
[junit4] 2> 3462589 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy
http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/ has no replicas
[junit4] 2> 3462589 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in
election, clear LIR
[junit4] 2> 3462593 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21]
o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/ shard1
[junit4] 2> 3462696 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ZkController
I am the leader, no recovery necessary
[junit4] 2> 3462699 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz
c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1375
[junit4] 2> 3462702 INFO (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz
c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null
path=/admin/collections
params={node=127.0.0.1:44392_pq_%252Fz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2}
status=0 QTime=1402
[junit4] 2> 3462799 INFO (zkCallback-10545-thread-1) [ ]
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent
state:SyncConnected type:NodeDataChanged
path:/collections/collection1/state.json] for collection [collection1] has
occurred - updating... (live nodes size: [2])
[junit4] 2> 3462882 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001
of type NRT
[junit4] 2> 3462883 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git:
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
[junit4] 2> 3462889 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 3462889 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 3462889 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 3462889 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@1aaa752d{/pq_/z,null,AVAILABLE}
[junit4] 2> 3462889 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@70cb9b0f{HTTP/1.1,[http/1.1]}{127.0.0.1:35419}
[junit4] 2> 3462889 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.e.j.s.Server Started @3462946ms
[junit4] 2> 3462889 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_jetty2,
solrconfig=solrconfig.xml, hostContext=/pq_/z, hostPort=35419,
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001/cores}
[junit4] 2> 3462890 ERROR
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 3462891 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter Using logger factory
org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 3462891 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.7.0
[junit4] 2> 3462891 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 3462891 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 3462891 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2018-11-27T17:55:38.474Z
[junit4] 2> 3462893 INFO (zkConnectionManagerCallback-10556-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3462895 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 3462895 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001/solr.xml
[junit4] 2> 3462899 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay
is ignored
[junit4] 2> 3462899 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter
autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 3462901 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 3462904 DEBUG (ScheduledTrigger-12519-thread-3) [ ]
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with
currently live nodes: 2
[junit4] 2> 3463304 INFO
(OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
[n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000004 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 3463361 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr
[junit4] 2> 3463363 INFO (zkConnectionManagerCallback-10560-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3463365 INFO (zkConnectionManagerCallback-10562-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3463374 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 3463379 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 3463379 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:35419_pq_%2Fz
[junit4] 2> 3463381 INFO (zkCallback-10526-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 3463381 INFO (zkCallback-10537-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 3463381 INFO (zkCallback-10552-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 3463384 INFO (zkCallback-10545-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 3463384 INFO (zkCallback-10519-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 3463388 INFO (zkCallback-10561-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 3463409 INFO (zkConnectionManagerCallback-10569-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 3463411 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 3463412 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
[junit4] 2> 3463412 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history
in memory.
[junit4] 2> 3463445 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node')
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3463462 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm')
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3463462 INFO
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry
'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
[junit4] 2> 3463465 INFO (TEST-HdfsRestartWhileUpdatingTest.
[...truncated too long message...]
kout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml
resolve:
jar-checksums:
[mkdir] Created dir:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894
[copy] Copying 241 files to
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894
[delete] Deleting directory
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894
check-working-copy:
[ivy:cachepath] :: resolving dependencies ::
org.eclipse.jgit#org.eclipse.jgit-caller;working
[ivy:cachepath] confs: [default]
[ivy:cachepath] found
org.eclipse.jgit#org.eclipse.jgit;4.6.0.201612231935-r in public
[ivy:cachepath] found com.jcraft#jsch;0.1.53 in public
[ivy:cachepath] found com.googlecode.javaewah#JavaEWAH;1.1.6 in public
[ivy:cachepath] found org.apache.httpcomponents#httpclient;4.3.6 in
public
[ivy:cachepath] found org.apache.httpcomponents#httpcore;4.3.3 in public
[ivy:cachepath] found commons-logging#commons-logging;1.1.3 in public
[ivy:cachepath] found commons-codec#commons-codec;1.6 in public
[ivy:cachepath] found org.slf4j#slf4j-api;1.7.2 in public
[ivy:cachepath] :: resolution report :: resolve 29ms :: artifacts dl 2ms
---------------------------------------------------------------------
| | modules || artifacts |
| conf | number| search|dwnlded|evicted|| number|dwnlded|
---------------------------------------------------------------------
| default | 8 | 0 | 0 | 0 || 8 | 0 |
---------------------------------------------------------------------
[wc-checker] Initializing working copy...
[wc-checker] SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
[wc-checker] SLF4J: Defaulting to no-operation (NOP) logger implementation
[wc-checker] SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for
further details.
[wc-checker] Checking working copy status...
-jenkins-base:
BUILD SUCCESSFUL
Total time: 285 minutes 41 seconds
Archiving artifacts
java.lang.InterruptedException: no matches found within 10000
at hudson.FilePath$34.hasMatch(FilePath.java:2678)
at hudson.FilePath$34.invoke(FilePath.java:2557)
at hudson.FilePath$34.invoke(FilePath.java:2547)
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2918)
Also: hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene
at
hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
at
hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
at hudson.remoting.Channel.call(Channel.java:955)
at hudson.FilePath.act(FilePath.java:1036)
at hudson.FilePath.act(FilePath.java:1025)
at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
at
hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
at
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
at
hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
at
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
at hudson.model.Build$BuildExecution.post2(Build.java:186)
at
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
at hudson.model.Run.execute(Run.java:1819)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at
hudson.model.ResourceController.execute(ResourceController.java:97)
at hudson.model.Executor.run(Executor.java:429)
Caused: hudson.FilePath$TunneledInterruptedException
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2920)
at hudson.remoting.UserRequest.perform(UserRequest.java:212)
at hudson.remoting.UserRequest.perform(UserRequest.java:54)
at hudson.remoting.Request$2.run(Request.java:369)
at
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:748)
Caused: java.lang.InterruptedException: java.lang.InterruptedException: no
matches found within 10000
at hudson.FilePath.act(FilePath.java:1038)
at hudson.FilePath.act(FilePath.java:1025)
at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
at
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
at
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
at hudson.model.Build$BuildExecution.post2(Build.java:186)
at
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
at hudson.model.Run.execute(Run.java:1819)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:97)
at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern
"**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error?
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]