> Also in our favour is that the users of the cluster we are currently 
> intending for this have established a practice of storing large objects.

That definitely is in your favor.

> but it remains to be seen how 60x 22TB behaves in practice.

Be sure you don't get SMR drives.

>  and it's hard for it to rebalance.


^ This.

> What is OLC?

QLC SSDs store 33% more data than TLC, 4 voltage levels per cell vs 3.

> Fascinating to hear about destroy-redeploy being safer than a simple 
> restart-recover!

This was Luminous, that dynamic may be different now, esp. with Nautilus async 
recovery.  

> Agreed. I guess I wanted to add the data point that these kinds of clusters 
> can and do make full sense in certain contexts, and push a little away from 
> "friends don't let friends use HDDs" dogma.

Understood.  Some deployments aren't squeezed for DC space -- today.  But since 
many HDD deployments are using LFF chassis, the form factor and interface 
limitations down the road still complicate expansion and SSD utilization.

> For now, we limit individual cloud volumes to 300 IOPs, doubled for those who 
> need it.

I'm curious how many clients / volumes you have vs. number of HDD OSDs and if 
you're using replication or EC.  If you have relatively few clients per HDD 
that would definitely improve the dynamic.



_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to