[ceph-users] Re: OSDs crush - Since Pacific

2022-09-01 Thread Igor Fedotov
Hi Wissem, given the log output it looks like suicide timeout has been fired. From my experience this is often observed when DB performance is degraded after bulk removals. And offline compaction should provide some relief. At least temporarily... But if deletes are ongoing (e.g. due to cluste

[ceph-users] Re: OSDs crush - Since Pacific

2022-08-30 Thread Igor Fedotov
Hi Wissem, sharing OSD log snippet preceding the crash (e.g. prior 20K lines) could be helpful and hopefully will provide more insigh - there might be some errors/assertion details and/or other artefacts... Thanks, Igor On 8/30/2022 10:51 AM, Wissem MIMOUNA wrote: Hi Stefan, We don’t have

[ceph-users] Re: OSDs crush - Since Pacific

2022-08-30 Thread Wissem MIMOUNA
Hi Stefan, We don’t have automatic conversion going on , and the « bluestore_fsck_quick_fix_on_mount » is not set . So we did an offline compaction as suggested but this didn’t fix the problem os osd crush . In the meantime we are rebuilding all OSDs on the cluster and it seems it improve the cl