Hello all 

Need urgent help on below… 
I tried reducing min_size but still showing same… 

cluster:
    id:     15688cb4-044a-11ec-942e-516035adea04
    health: HEALTH_ERR
            3 failed cephadm daemon(s)
            1 filesystem is degraded
            1 MDSs report slow metadata IOs
            20/16670718 objects unfound (0.000%)
            Reduced data availability: 283 pgs inactive, 464 pgs incomplete
            Possible data damage: 2 pgs recovery_unfound
            Degraded data redundancy: 42998/61175329 objects degraded (0.070%), 
2 pgs degraded, 1 pg undersized
            304 pgs not deep-scrubbed in time
            1055 slow ops, oldest one blocked for 94042 sec, daemons 
[osd.1,osd.12,osd.13,osd.15,osd.16,osd.19,osd.20,osd.21,osd.26,osd.27]... have 
slow ops.

  services:
    mon:        5 daemons, quorum 
van2-converged04n,van2-converged05n,van2-converged01n,van2-converged03n,van2-converged02n
 (age 17h)
    mgr:        van2-converged05n.lybhho(active, since 22h), standbys: 
van2-converged04n.azmqik
    mds:        3/3 daemons up, 11 standby
    osd:        31 osds: 31 up (since 5h), 31 in (since 5h); 2 remapped pgs
    rbd-mirror: 2 daemons active (2 hosts)

  data:
    volumes: 1/2 healthy, 1 recovering
    pools:   32 pools, 1505 pgs
    objects: 16.67M objects, 56 TiB
    usage:   152 TiB used, 234 TiB / 387 TiB avail
    pgs:     0.332% pgs unknown
             30.897% pgs not active
             42998/61175329 objects degraded (0.070%)
             21469/61175329 objects misplaced (0.035%)
             20/16670718 objects unfound (0.000%)
             1034 active+clean
             463  incomplete
             5    unknown
             1    remapped+incomplete
             1    active+recovery_unfound+degraded
             1    recovery_unfound+undersized+degraded+remapped+peered

  progress:
    Global Recovery Event (22h)
      [===================.........] (remaining: 10h)


[WRN] SLOW_OPS: 1055 slow ops, oldest one blocked for 93877 sec, daemons 
[osd.1,osd.12,osd.13,osd.15,osd.16,osd.19,osd.20,osd.21,osd.26,osd.27]... have 
slow ops.
[WRN] MDS_SLOW_METADATA_IO: 1 MDSs report slow metadata IOs
    mds.van2.van2-converged05n.mbbzfj(mds.1): 1 slow metadata IOs are blocked > 
30 secs, oldest blocked for 62279 secs
[WRN] OBJECT_UNFOUND: 20/16670718 objects unfound (0.000%)
    pg 28.1f has 9 unfound objects
    pg 29.2 has 11 unfound objects
[WRN] PG_AVAILABILITY: Reduced data availability: 470 pgs inactive, 464 pgs 
incomplete
    pg 16.58 is incomplete, acting [7,14,19] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.5c is incomplete, acting [26,3,19] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.5f is incomplete, acting [13,6,21] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.63 is incomplete, acting [9,7,15] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.64 is incomplete, acting [18,19,5] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.68 is incomplete, acting [2,5,27] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.70 is incomplete, acting [19,1,24] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.74 is incomplete, acting [8,25,14] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.78 is incomplete, acting [23,14,29] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.7c is incomplete, acting [24,10,29] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 16.7d is incomplete, acting [12,22,13] (reducing pool lv-r3-for-ec-disks 
min_size from 2 may help; search ceph.com/docs for 'incomplete')
    pg 17.64 is incomplete, acting [20,11,13] (reducing pool 
lv-r3-for-ec-large-disks min_size from 2 may help; search ceph.com/docs for 
'incomplete')
    pg 17.66 is incomplete, acting [21,13,27] (reducing pool 
lv-r3-for-ec-large-disks min_size from 2 may help; search ceph.com/docs for 
'incomplete')
    pg 17.6a is incomplete, acting [4,26,13] (reducing pool 
lv-r3-for-ec-large-disks min_size from 2 may help; search ceph.com/docs for 
'incomplete')
    pg 17.6c is incomplete, acting [13,12,31] (reducing pool 
lv-r3-for-ec-large-disks min_size from 2 may help; search ceph.com/docs for 
'incomplete')
    pg 17.6e is incomplete, acting [8,31,24] (reducing pool 
lv-r3-for-ec-large-disks min_size from 2 may help; search ceph.com/docs for 
'incomplete')
    pg 17.7a is incomplete, acting [8,28,27] (reducing pool 
lv-r3-for-ec-large-disks min_size from 2 may help; search ceph.com/docs for 
'incomplete')
    pg 19.54 is stuck inactive since forever, current state incomplete, last 
acting [10,4,20,15,11] (reducing pool lvp-ec-large-disks min_size from 3 may 
help; search ceph.com/docs for 'incomplete')
    pg 19.58 is incomplete, acting [15,18,12,28,20] (reducing pool 
lvp-ec-large-disks min_size from 3 may help; search ceph.com/docs for 
'incomplete')
    pg 19.59 is incomplete, acting [6,9,18,21,26] (reducing pool 
lvp-ec-large-disks min_size from 3 may help; search ceph.com/docs for 
'incomplete')
    pg 19.5a is incomplete, acting [11,31,20,17,24] (reducing pool 
lvp-ec-large-disks min_size from 3 may help; search ceph.com/docs for 
‘incomplceph -s
Ceph 

Any way to make these scrubs complete faster?

pg 19.63 not deep-scrubbed since 2025-05-24T08:18:29.738427+0000
    pg 34.6d not deep-scrubbed since 2025-06-04T08:33:54.534882+0000
    pg 16.5f not deep-scrubbed since 2025-05-21T10:44:09.996254+0000
    pg 19.5d not deep-scrubbed since 2025-05-26T09:36:27.064154+0000
    pg 19.5e not deep-scrubbed since 2025-06-05T00:52:03.859984+0000
    pg 16.5c not deep-scrubbed since 2025-06-06T00:36:22.021390+0000
    pg 19.5f not deep-scrubbed since 2025-06-03T16:27:42.356213+0000
    pg 22.5a not deep-scrubbed since 2025-06-05T23:00:28.066065+0000
    pg 34.69 not deep-scrubbed since 2025-06-03T05:07:58.209808+0000
    pg 19.58 not deep-scrubbed since 2025-05-27T23:32:29.963976+0000
    pg 19.59 not deep-scrubbed since 2025-05-25T11:50:44.735318+0000
    pg 19.5a not deep-scrubbed since 2025-06-06T02:34:05.486126+0000
    pg 16.58 not deep-scrubbed since 2025-05-13T14:10:44.570493+0000


Regards
Dev
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to