> On Apr 25, 2018, at 7:49 AM, Allen Wittenauer <a...@effectivemachines.com> > wrote: > >> Using the yetus jobs as a reference, yetus-java builds 480 and 481 are >> nearly a year old, but only contain a few kilobytes of data. While removing >> them saves no space, they also provide no value, > > … to infra. > > The value to the communities that any job services is really up to > those communities to decide.
I mean builds 480 and 481 literally provide no value, since they are not accessible via the jenkins UI and only contain a polling.log file from 2017. I recognize that some projects may want to retain specific older builds, but I personally question the utility of data kept older than 6 months. For build data that’s significantly outside the nominal 30 day / 10 job window, we’d ask that this be exported and managed locally by the project rather than remaining “live” on the master. Is there a reason other than convenience for it to remain live? Just based on some initial review, excluding the build logs and job metadata is probably do-able for an initial pass at purging old data, but I’ll want to generate some data on how many old build histories exist. As I stated earlier the main goal here is to remove dead jobs and binary artifacts. There do appear to be a fair few jobs which no longer exist as mentioned else-thread, so hopefully we’ll realize some notable performance and space improvements by culling the low hanging fruit before a more drastic approach is required. -Chris
signature.asc
Description: Message signed with OpenPGP