On Mon, Feb 13, 2023 at 4:16 AM Sake Paulusma <sake1...@hotmail.com> wrote:
>
> Hello,
>
> I configured a stretched cluster on two datacenters. It's working fine, 
> except this weekend the Raw Capicity exceeded 50% and the error 
> POOL_TARGET_SIZE_BYTES_OVERCOMMITED showed up.
>
> The command "ceph df" is showing the correct cluster size, but "ceph osd pool 
> autoscale-status" is showing half of the total Raw Capacity.
>
> What could be wrong?

There's a bug with the statistics handling of pools in stretch mode,
and others like them. :(
https://tracker.ceph.com/issues/56650

-Greg


>
>
>
> ----------------------------------------------------------------------------
> [ceph: root@aqsoel11445 /]# ceph status
>   cluster:
>     id:     adbe7bb6-5h6d-11ed-8511-004449ede0c
>     health: HEALTH_WARN
>             1 MDSs report oversized cache
>             1 subtrees have overcommitted pool target_size_bytes
>
>   services:
>     mon: 5 daemons, quorum host1,host2,host3,host4,host5 (age 4w)
>     mgr: aqsoel11445.nqamuz(active, since 5w), standbys: host1.wujgas
>     mds: 2/2 daemons up, 2 standby
>     osd: 12 osds: 12 up (since 5w), 12 in (since 9w)
>
>   data:
>     volumes: 2/2 healthy
>     pools:   5 pools, 193 pgs
>     objects: 17.31M objects, 1.2 TiB
>     usage:   5.0 TiB used, 3.8 TiB / 8.8 TiB avail
>     pgs:     192 active+clean
>              1   active+clean+scrubbing
> ----------------------------------------------------------------------------
>
> ----------------------------------------------------------------------------
> [ceph: root@aqsoel11445 /]# ceph df
> --- RAW STORAGE ---
> CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED
> ssd    8.8 TiB  3.8 TiB  5.0 TiB   5.0 TiB      56.83
> TOTAL  8.8 TiB  3.8 TiB  5.0 TiB   5.0 TiB      56.83
>
> --- POOLS ---
> POOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX 
> AVAIL
> .mgr                        1    1  449 KiB        2  1.8 MiB      0    320 
> GiB
> cephfs.application-tst.meta   2   16  540 MiB   18.79k  2.1 GiB   0.16    320 
> GiB
> cephfs.application-tst.data   3   32  4.4 GiB    8.01k   17 GiB   1.33    320 
> GiB
> cephfs.application-acc.meta   4   16   11 GiB    3.54M   45 GiB   3.37    320 
> GiB
> cephfs.application-acc.data   5  128  1.2 TiB   13.74M  4.8 TiB  79.46    320 
> GiB
> ----------------------------------------------------------------------------
>
> ----------------------------------------------------------------------------
> [ceph: root@aqsoel11445 /]# ceph osd pool autoscale-status
> POOL                         SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO  
> TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE  BULK
> .mgr                       448.5k                4.0         4499G  0.0000    
>                               1.0       1              on         False
> cephfs.application-tst.meta  539.8M                4.0         4499G  0.0005  
>                                 4.0      16              on         False
> cephfs.application-tst.data   4488M       51200M   4.0         4499G  0.0444  
>                                 1.0      32              on         False
> cephfs.application-acc.meta  11430M                4.0         4499G  0.0099  
>                                 4.0      16              on         False
> cephfs.application-acc.data   1244G                4.0         4499G  1.1062  
>       1.0000           0.9556   1.0     128              on         False
> ----------------------------------------------------------------------------
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to