On Mon, 19 Feb 2024 at 08:59, Kevin Kofler via devel < devel@lists.fedoraproject.org> wrote:
> Miroslav Suchý wrote: > > What **you** would find as acceptable policy for pruning rawhide chroots? > > As I mentioned several times, I already find the existing policy for > pruning > EOL release chroots unacceptable (because deleting data must never be the > default – notifications can be and are still lost in spam filters, I still > do not ever get any notification from Copr! – and because the UI to extend > the lifetime follows dark patterns, requiring us to click separately for > every single chroot instead of having an "Extend all" button). > > Instead of coming up with new aggressive pruning schemes, Copr really > needs > to come up with a reasonable amount of storage to satisfy user demands. > HDDs > in the multi-TB-range are available for fairly low budgets (extremely low > by > the standards of a company like IBM), and it just takes 2 of them to build > a > RAID 1 that is safe against data loss. Of course, that means that Copr > needs > to stop locking itself into third-party cloud providers that charge > ridiculously high prices for storage. > Kevin, I agree with your general concerns but have problems with your analysis of costs. 1. Drive size is not just what is needed but also throughput. The large drives needed to store the data COPR uses for its hundreds of chroots are much 'slower' on reads and writes even when adding in layers of RAID 1+0. Faster drives are possible but the price goes up considerably. 2. Throughput of individual drives also requires backplane speeds which match peek throughput of all the drives. Otherwise you end up with lots of weird stalling (as seen on certain builders which have such drives). 3. Outside of that you need to have a fast network switch speed of 10G minimum and 40G to 100G to deal with the multiple builders and the storage server. The larger the storage, the more bandwidth needed all the way through as building software eats lots of space. 4. The builders need to be housed in a datacenter which charges for a. power b. cooling c. square footage used d. staff to deal with problems. e. racks and wiring and parts Going from the costs 5 years ago, it took using storage systems from 45drives and systems elsewhere.. The capital costs for COPR was around 350k. The operating expenses were around 80k per year. At the time, chroots and other things needed to be cleaned much more regularly than now. We would probably need to double or triple those costs to meet what we have now. [budgets like what was given then only came around 1/10 years or so.] The amazon systems cost Fedora $0.00 as long as it stays within 'modest' disk space and other resources. That said, I realize there will be a day when you can clearly said "I told you this would happen" when being locked in comes back to bite. -- Stephen Smoogen, Red Hat Automotive Let us be kind to one another, for most of us are fighting a hard battle. -- Ian MacClaren
-- _______________________________________________ devel mailing list -- devel@lists.fedoraproject.org To unsubscribe send an email to devel-le...@lists.fedoraproject.org Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/ List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines List Archives: https://lists.fedoraproject.org/archives/list/devel@lists.fedoraproject.org Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue