Hey Greg,

I'm just analyzing this issue and it isn't strange the total cluster size is 
half the total size (or the smallest of both clusters). Because you shouldn't 
write more data to the cluster than the smallest datacenter can handle.
Second when in datacenter fail over modus, the cluster size

________________________________
From: Gregory Farnum <gfar...@redhat.com>
Sent: Monday, February 13, 2023 5:32:18 PM
To: Sake Paulusma <sake1...@hotmail.com>
Cc: ceph-users@ceph.io <ceph-users@ceph.io>
Subject: Re: [ceph-users] Health warning - POOL_TARGET_SIZE_BYTES_OVERCOMMITED

On Mon, Feb 13, 2023 at 4:16 AM Sake Paulusma <sake1...@hotmail.com> wrote:
>
> Hello,
>
> I configured a stretched cluster on two datacenters. It's working fine, 
> except this weekend the Raw Capicity exceeded 50% and the error 
> POOL_TARGET_SIZE_BYTES_OVERCOMMITED showed up.
>
> The command "ceph df" is showing the correct cluster size, but "ceph osd pool 
> autoscale-status" is showing half of the total Raw Capacity.
>
> What could be wrong?

There's a bug with the statistics handling of pools in stretch mode,
and others like them. :(
https://tracker.ceph.com/issues/56650

-Greg


>
>
>
> ----------------------------------------------------------------------------
> [ceph: root@aqsoel11445 /]# ceph status
>   cluster:
>     id:     adbe7bb6-5h6d-11ed-8511-004449ede0c
>     health: HEALTH_WARN
>             1 MDSs report oversized cache
>             1 subtrees have overcommitted pool target_size_bytes
>
>   services:
>     mon: 5 daemons, quorum host1,host2,host3,host4,host5 (age 4w)
>     mgr: aqsoel11445.nqamuz(active, since 5w), standbys: host1.wujgas
>     mds: 2/2 daemons up, 2 standby
>     osd: 12 osds: 12 up (since 5w), 12 in (since 9w)
>
>   data:
>     volumes: 2/2 healthy
>     pools:   5 pools, 193 pgs
>     objects: 17.31M objects, 1.2 TiB
>     usage:   5.0 TiB used, 3.8 TiB / 8.8 TiB avail
>     pgs:     192 active+clean
>              1   active+clean+scrubbing
> ----------------------------------------------------------------------------
>
> ----------------------------------------------------------------------------
> [ceph: root@aqsoel11445 /]# ceph df
> --- RAW STORAGE ---
> CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED
> ssd    8.8 TiB  3.8 TiB  5.0 TiB   5.0 TiB      56.83
> TOTAL  8.8 TiB  3.8 TiB  5.0 TiB   5.0 TiB      56.83
>
> --- POOLS ---
> POOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX 
> AVAIL
> .mgr                        1    1  449 KiB        2  1.8 MiB      0    320 
> GiB
> cephfs.application-tst.meta   2   16  540 MiB   18.79k  2.1 GiB   0.16    320 
> GiB
> cephfs.application-tst.data   3   32  4.4 GiB    8.01k   17 GiB   1.33    320 
> GiB
> cephfs.application-acc.meta   4   16   11 GiB    3.54M   45 GiB   3.37    320 
> GiB
> cephfs.application-acc.data   5  128  1.2 TiB   13.74M  4.8 TiB  79.46    320 
> GiB
> ----------------------------------------------------------------------------
>
> ----------------------------------------------------------------------------
> [ceph: root@aqsoel11445 /]# ceph osd pool autoscale-status
> POOL                         SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO  
> TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE  BULK
> .mgr                       448.5k                4.0         4499G  0.0000    
>                               1.0       1              on         False
> cephfs.application-tst.meta  539.8M                4.0         4499G  0.0005  
>                                 4.0      16              on         False
> cephfs.application-tst.data   4488M       51200M   4.0         4499G  0.0444  
>                                 1.0      32              on         False
> cephfs.application-acc.meta  11430M                4.0         4499G  0.0099  
>                                 4.0      16              on         False
> cephfs.application-acc.data   1244G                4.0         4499G  1.1062  
>       1.0000           0.9556   1.0     128              on         False
> ----------------------------------------------------------------------------
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to