I should have mentioned that all 10 of the jenkins.util.Timer's are in the 
RUNNABLE state and the Jenkins cron thread is in TIMED_WAITING which seems 
to make sense (I think).

The queued and completed tasks from jenkins.util.Timer.get() hasn't changed 
appreciably
On Friday, January 15, 2016 at 9:00:23 AM UTC-5, Kevin wrote:
>
> Wow, can't believe I didn't know about /threadDump before, thanks!
>
> Anyways I mostly see lots (on the order of 100s) of:
>
> "Computer.threadPoolForRemoting [#85811]" Id=380625 Group=main 
> TIMED_WAITING on 
> java.util.concurrent.SynchronousQueue$TransferStack@687826ae
> at sun.misc.Unsafe.park(Native Method)
> -  waiting on java.util.concurrent.SynchronousQueue$TransferStack@687826ae
> at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> at 
> java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460)
> at 
> java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362)
> at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941)
> at 
> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1066)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1127)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
>
>
> I did see some (~20) CPS executors that looked suspicious (they're also 
> hanging forever when i go to the job):
>
> "Running CpsFlowExecution[Owner[workflow_test/619:workflow_test#619]]" 
> Id=7310 Group=main RUNNABLE
> at 
> org.kohsuke.stapler.framework.io.LargeText$BufferSession.skip(LargeText.java:532)
> at 
> org.kohsuke.stapler.framework.io.LargeText.writeLogTo(LargeText.java:211)
> at 
> hudson.console.AnnotatedLargeText.writeRawLogTo(AnnotatedLargeText.java:162)
> at 
> org.jenkinsci.plugins.workflow.job.WorkflowRun.copyLogs(WorkflowRun.java:359)
> at 
> org.jenkinsci.plugins.workflow.job.WorkflowRun.access$600(WorkflowRun.java:107)
> at 
> org.jenkinsci.plugins.workflow.job.WorkflowRun$GraphL.onNewHead(WorkflowRun.java:752)
> -  locked java.util.concurrent.atomic.AtomicBoolean@7eb2df49
> at 
> org.jenkinsci.plugins.workflow.cps.CpsFlowExecution.notifyListeners(CpsFlowExecution.java:799)
> at 
> org.jenkinsci.plugins.workflow.cps.CpsThreadGroup$4.run(CpsThreadGroup.java:320)
> at 
> org.jenkinsci.plugins.workflow.cps.CpsVmExecutorService$1.run(CpsVmExecutorService.java:32)
> at 
> hudson.remoting.SingleLaneExecutorService$1.run(SingleLaneExecutorService.java:112)
> at 
> jenkins.util.ContextResettingExecutorService$1.run(ContextResettingExecutorService.java:28)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
>
> Number of locked synchronizers = 1
> - java.util.concurrent.ThreadPoolExecutor$Worker@d3b1e0f
>
>
> I noticed them before but I assumed they weren't related to this, issue, I 
> think they're waiting for slaves nodes that have been removed from our 
> jenkins (not just disconnected but deleted), I'll clean these up with the 
> /kill url on workflows and see if that improves things since the threaddump 
> makes me wonder if these are related.
>
>
> Thanks again
>
> On Thursday, January 14, 2016 at 4:14:16 PM UTC-5, Christopher Orr wrote:
>>
>> On 14/01/16 21:30, Kevin wrote: 
>> > Hi all, I've got some custom cloud providers that don't seem to get 
>> > triggered, I've traced it back to NodeProvisionerInvoker not being 
>> > called (even after 30 minutes of waiting), I can call the 
>> > nodeProvisioners by hand (using the update function) in the same way 
>> > that the NodeProvisionerInvoker does in the groovy script console and 
>> > things seem to work.  Actually it seems like a number of bits related 
>> to 
>> > PeriodicWork classes aren't working since I assume git polling and the 
>> > periodic build setting are a PeriodicWork things, but they're also not 
>> > triggering. 
>> > 
>> > I did also look at PeriodicWork.all().collect {it.class.name} 
>> > and println jenkins.util.Timer.get() and got: 
>> > 
>> >     hudson.diagnosis.HudsonHomeDiskUsageChecker, 
>> >     hudson.diagnosis.MemoryUsageMonitor, 
>> >     hudson.model.FingerprintCleanupThread, 
>> >     hudson.model.LoadStatistics$LoadStatisticsUpdater, 
>> >     hudson.model.WorkspaceCleanupThread, 
>> >     hudson.slaves.ComputerRetentionWork, 
>> >     hudson.slaves.ConnectionActivityMonitor, 
>> >     hudson.slaves.NodeProvisioner$NodeProvisionerInvoker, 
>> >     hudson.triggers.Trigger$Cron, 
>> >     jenkins.model.DownloadSettings$DailyCheck, 
>> >     org.jenkinsci.plugins.periodicreincarnation.PeriodicReincarnation] 
>> > 
>> > 
>> > and 
>> > 
>> >     
>> jenkins.util.ErrorLoggingScheduledThreadPoolExecutor@6d72bb0e[Running, 
>> >     pool size = 10, active threads = 10, queued tasks = 2521, completed 
>> >     tasks = 1134830] 
>>
>> If the number of active threads remains equal to the pool size, then I 
>> would guess that some tasks are getting stuck. 
>>
>> Can you see a bunch of timer-related or otherwise suspicious-looking 
>> tasks running if you go to /threadDump on your Jenkins instance? 
>>
>> Regards, 
>> Chris 
>>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Jenkins Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to jenkinsci-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/jenkinsci-users/754475d6-b63e-4e21-9fa5-a211e870129a%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to