We have a build pipeline that dynamically scales up agents (with K8s 
plugin) to execute tests on. We've noticed that as the amount of agents 
scales up to 20 (or even over 50), the performance of each agent degrades. 

Each agent's resource usage is relatively low, but the requests are done in 
a way that at most 2 agents end up on a single k8s node. And based on the 
metrics collected from prometheus, there is no indication that the pods 
themselves are under high load. We have seen the Jenkins master hit high 
CPU levels (7 core), but this happens more often when concurrent builds 
with high agent counts.

In fact, I've added `time` to each `sh` command in the pipeline. And so for 
example, a `yarn install --ignore-optional` (with packages already 
installed) takes approx 2 seconds, but Jenkins is reporting it as 1 minute 
28 seconds. (see screenshot)

Is there anything that can be done/adjusted to resolve this?

/Jeff 

-- 
You received this message because you are subscribed to the Google Groups 
"Jenkins Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to jenkinsci-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/jenkinsci-users/3d0c580f-4b15-4210-ba8f-d9d1b2ae6776n%40googlegroups.com.

Reply via email to