Hi all,

I am testing rgw cluster which separated between index pool osd and data
pool osd.
After uploading a lot of objects, I found that index pool osd usage is
unbalanced. I just know that index pool use rocksdb. rocksdb has object
metadata, bucket metadata, multipart, versioning data for index pool.
I thought that most of usage is object metadata. so I checked balance state
of object metadata between osds of index pool. it's balanced well. (I know
it though OMAP stat).
Which part can make unbalance between index pool disks?

(you can check osd.27, osd.29 which are bigger than other osds)
ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META
 AVAIL    %USE   VAR   PGS  STATUS
24    ssd  3.49309   1.00000  3.5 TiB  1.8 TiB  399 MiB  1.2 TiB   581 GiB
 1.7 TiB  51.72  0.85   71      up
25    ssd  3.49309   1.00000  3.5 TiB  2.2 TiB  179 MiB  1.4 TiB   836 GiB
 1.3 TiB  62.22  1.02   71      up
26    ssd  3.49309   1.00000  3.5 TiB  2.0 TiB  180 MiB  1.2 TiB   752 GiB
 1.5 TiB  56.55  0.93   72      up
*27    ssd  3.49309   1.00000  3.5 TiB  2.7 TiB  399 MiB  1.3 TiB   1.4 TiB
 815 GiB  77.22  1.27   70      up*
28    ssd  3.49309   1.00000  3.5 TiB  1.5 TiB  179 MiB  1.2 TiB   321 GiB
 2.0 TiB  43.57  0.71   72      up
*29    ssd  3.49309   1.00000  3.5 TiB  2.8 TiB  179 MiB  1.4 TiB   1.4 TiB
 748 GiB  79.08  1.30   73      up*
30    ssd  3.49309   1.00000  3.5 TiB  1.7 TiB  179 MiB  1.4 TiB   342 GiB
 1.8 TiB  49.37  0.81   75      up
31    ssd  3.49309   1.00000  3.5 TiB  2.5 TiB  179 MiB  1.4 TiB   1.2 TiB
 969 GiB  72.90  1.20   69      up
32    ssd  3.49309   1.00000  3.5 TiB  2.4 TiB  179 MiB  1.3 TiB   1.1 TiB
 1.1 TiB  67.46  1.11   66      up
33    ssd  3.49309   1.00000  3.5 TiB  2.1 TiB  179 MiB  1.2 TiB  1015 GiB
 1.3 TiB  61.40  1.01   68      up
.... omission ....
                       TOTAL   63 TiB   38 TiB  3.8 GiB   23 TiB    16 TiB
  25 TiB  60.96
MIN/MAX VAR: 0.68/1.30  STDDEV: 9.98
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to