Hi,
On 2022-11-19 17:32, Anthony D'Atri wrote:
I’m not positive that the options work with hyphens in them. Try
ceph tell osd.* injectargs '--osd_max_backfills 1
--osd_recovery_max_active 1 --osd_recovery_max_single_start 1
--osd_recovery_op_priority=1'
Did so.
With Quincy the following should already be set, but to be sure:
ceph tell osd.* config set osd_op_queue_cut_off high
Did so too and even restarted all osd as it was recommended.
I then stopped a single osd in order to cause some backfilling.
What is network saturation like on that 1GE replication network?
Typically 100% saturated.
Operations like yours that cause massive data movement could easily
saturate a pipe that narrow.
Sure, but I am used to other setups where the recovery can be slowed
down in order to keep the rbds operating.
To me it looks like all backfilling happens in parallel without any
pauses in between which would benefit the client traffic.
I would expect some of those pgs in
active+undersized+degraded+remapped+backfill_wait state instead of
backfilling.
2022-11-19T16:58:50.139390+0000 mgr.pve-02 (mgr.18134134) 61735 :
cluster [DBG] pgmap v60978: 576 pgs: 102
active+undersized+degraded+remapped+backfilling, 474 active+clean; 2.4
TiB data, 4.3 TiB used, 10 TiB / 15 TiB avail; 150 KiB/s wr, 10 op/s;
123337/1272524 objects degraded (9.692%); 228 MiB/s, 58 objects/s
recovering
Is this Quincy specific?
Regards
--martin
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io