Hi,
are you using mclock scheduler (default in Quincy)? Until Reef 18.2.4
there was a default value set for osd_snap_trim_cost (1M bytes) which
blocked snaptrims [0]. This was fixed in [1] and backported to Reef.
But it's unlikely that this was your issue in Octopus as mclock became
the default in Pacific, IIRC. Since Quincy is also EOL, I'd recommed
to update further, if possible.
Were you able to avoid OSD flapping with the nodown flag (ceph osd set
nodown)? This can help to keep the cluster more stable in such
situations.
Can you add some more details about your setup like:
ceph -s
ceph osd df tree
ceph osd pool ls detail
ceph df
Are you using HDD OSDs or HDDs with dedicated DB/WAL? How many
snapshots are you generating?
Regards,
Eugen
[0] https://tracker.ceph.com/issues/67702
[1] https://tracker.ceph.com/issues/63604
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]