Hi,

We upgraded one of our production cluster (480 OSDs, 13,5 K PGs, most pools EC 9+6) from 18.2.7 to 19.2.2 on May 26. It was healthy when we upgraded it and remained so until last Thrusday (June 5) where, doing a `ceph -s`, I saw that there were 1 deep scrub and ~1000 scrubs late:

            1 pgs not deep-scrubbed in time
            1168 pgs not scrubbed in time

Looking again Friday morning (~16 hours later), this number increased a lot:

            27 pgs not deep-scrubbed in time
            3252 pgs not scrubbed in time

Checking our configuration on Friday, we found that osd_max_scrubs was set to 1 instead of using the new default since Reef which is 3 (probably a leftover of a config change after a problem 18 monts ago). We unset the specific value and this led to a reduction of these numbers in the next 24 hours (~2700 scrubs late) but since then (2 days) it remained stable initially and is now increasing slowly. This morning the situation is:

            294 pgs not deep-scrubbed in time
            3013 pgs not scrubbed in time

My guess is that the real issue is the late scrubs that results in many OSD reaching the limit of 3 concurrent scrubs and that it has the consequences that some deep scrubs cannot run too (I've in mind that the limit applies both to shallow scubs and deep scrubs, am I right?). I checked the main cluster logs  and didn't find any error or warning, related to OSDs like slow ops, slow requests...The only things we have spotted through our monitoring system is a dramatic decrease (~75%) of IOPS on each OSD server right after the 19.2.2 upgrade but it is not necessarily the sign of a problem. I guess it may in particular be a consequence of the new deep scrubs parameters, osd_shallow_scrub_chunk_min/max, which a probably intended to reduced the deep scrub IOPS load. The release notes for Squid also mention a change in osd_op_num_shards_hdd and osd_op_num_threads_per_shard_hdd, I don't know if they may also have impact.

Up to now, no users has reported any issue so it seems to be a problem only with scrubs. I'm wondering where to start looking for an issue or anything related to 19.2.2 is already known. We increased the deep scrub interval from 10 to 14 days  a few days before the upgrade (we saw that there was permanently 1 deep scrub late, a different PG all the time) and kept the standard 7 day interval for scrubs. Looking at the number of scrubs and deep scrubs per day, it doesn't look weird (see below).

I guess that if we retart all OSDs we'll clear the problem but we'd like to understand what happened and be sure it is not something related to Squid, before upgrade our other production cluster. Any hint/advice will be highly appreciated. I took a snapshot of `ceph pg dump pgs_brief` regularly and I'll try to identify if there are some stucked scrubs and what are the OSDs involved but with 500 OSDs and 18 OSD servers, it may not be obvious...

Best regards,

Michel

_Distribution of scrubs_ (first number is the number of scrubs during the day)

(number increases on June 6 after setting osd_max_scrubs=3)

    295 "2025-05-25
   1578 "2025-05-26
    300 "2025-05-27
    392 "2025-05-28
    578 "2025-05-29
    707 "2025-05-30
    611 "2025-05-31
    819 "2025-06-01
    679 "2025-06-02
    724 "2025-06-03
    698 "2025-06-04
    726 "2025-06-05
   1577 "2025-06-06
   1393 "2025-06-07
   1962 "2025-06-08
    645 "2025-06-09

_Distribution of deep scrubs per day_

(number increases on June 6 after setting osd_max_scrubs=3 and starts to decrease again on June 8, when the number of late subs increases again, probably because we hit the limit of 3 scrubs per OSD)

     22 "2025-05-12
     63 "2025-05-13
    101 "2025-05-14
    127 "2025-05-15
    173 "2025-05-16
    238 "2025-05-17
    305 "2025-05-18
    387 "2025-05-19
    450 "2025-05-20
    564 "2025-05-21
    675 "2025-05-22
    716 "2025-05-23
    871 "2025-05-24
   1071 "2025-05-25
    801 "2025-05-26
    188 "2025-05-27
    292 "2025-05-28
    335 "2025-05-29
    409 "2025-05-30
    371 "2025-05-31
    514 "2025-06-01
    440 "2025-06-02
    504 "2025-06-03
    478 "2025-06-04
    546 "2025-06-05
   1132 "2025-06-06
   1022 "2025-06-07
    662 "2025-06-08
    227 "2025-06-09
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to