Hi,

are you using mclock scheduler (default in Quincy)? Until Reef 18.2.4 there was a default value set for osd_snap_trim_cost (1M bytes) which blocked snaptrims [0]. This was fixed in [1] and backported to Reef. But it's unlikely that this was your issue in Octopus as mclock became the default in Pacific, IIRC. Since Quincy is also EOL, I'd recommed to update further, if possible. Were you able to avoid OSD flapping with the nodown flag (ceph osd set nodown)? This can help to keep the cluster more stable in such situations.
Can you add some more details about your setup like:

ceph -s
ceph osd df tree
ceph osd pool ls detail
ceph df

Are you using HDD OSDs or HDDs with dedicated DB/WAL? How many snapshots are you generating?

Regards,
Eugen

[0] https://tracker.ceph.com/issues/67702
[1] https://tracker.ceph.com/issues/63604
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to