> But not, I suspect, nearly as many tentacles.

No, that's the really annoying part. It just works.
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14

________________________________________
From: Anthony D'Atri <anthony.da...@gmail.com>
Sent: Thursday, October 10, 2024 2:13 PM
To: Frank Schilder
Cc: Peter Grandi; list Linux fs Ceph
Subject: Re: [ceph-users] What is the problem with many PGs per OSD



I'm afraid nobody will build a 100PB cluster with 1T drives. That's just absurd

Check the archives for the panoply of absurdity that I’ve encountered ;)

So, the sharp increase of per-device capacity has to be taken into account. 
Specifically as the same development is happening with SSDs. There is no way 
around 100TB drives in the near future and a system like ceph is either able to 
handle that or will die

Agreed.  I expect 122TB QLC in 1H2025.  With NVMe and PCI-e Gen 5 one might 
experiment with slicing each into two OSDs.  But for archival and object 
workloads latency usually isn’t so big a deal, so we may increasingly see a 
strategy adapted to the workloads.

10 higher aggregated sustained IOP/s performance compared with a similarly 
sized ceph cluster

But not, I suspect, nearly as many tentacles.



_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to