Hi all, I am testing rgw cluster which separated between index pool osd and data pool osd. After uploading a lot of objects, I found that index pool osd usage is unbalanced. I just know that index pool use rocksdb. rocksdb has object metadata, bucket metadata, multipart, versioning data for index pool. I thought that most of usage is object metadata. so I checked balance state of object metadata between osds of index pool. it's balanced well. (I know it though OMAP stat). Which part can make unbalance between index pool disks?
(you can check osd.27, osd.29 which are bigger than other osds) ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL %USE VAR PGS STATUS 24 ssd 3.49309 1.00000 3.5 TiB 1.8 TiB 399 MiB 1.2 TiB 581 GiB 1.7 TiB 51.72 0.85 71 up 25 ssd 3.49309 1.00000 3.5 TiB 2.2 TiB 179 MiB 1.4 TiB 836 GiB 1.3 TiB 62.22 1.02 71 up 26 ssd 3.49309 1.00000 3.5 TiB 2.0 TiB 180 MiB 1.2 TiB 752 GiB 1.5 TiB 56.55 0.93 72 up *27 ssd 3.49309 1.00000 3.5 TiB 2.7 TiB 399 MiB 1.3 TiB 1.4 TiB 815 GiB 77.22 1.27 70 up* 28 ssd 3.49309 1.00000 3.5 TiB 1.5 TiB 179 MiB 1.2 TiB 321 GiB 2.0 TiB 43.57 0.71 72 up *29 ssd 3.49309 1.00000 3.5 TiB 2.8 TiB 179 MiB 1.4 TiB 1.4 TiB 748 GiB 79.08 1.30 73 up* 30 ssd 3.49309 1.00000 3.5 TiB 1.7 TiB 179 MiB 1.4 TiB 342 GiB 1.8 TiB 49.37 0.81 75 up 31 ssd 3.49309 1.00000 3.5 TiB 2.5 TiB 179 MiB 1.4 TiB 1.2 TiB 969 GiB 72.90 1.20 69 up 32 ssd 3.49309 1.00000 3.5 TiB 2.4 TiB 179 MiB 1.3 TiB 1.1 TiB 1.1 TiB 67.46 1.11 66 up 33 ssd 3.49309 1.00000 3.5 TiB 2.1 TiB 179 MiB 1.2 TiB 1015 GiB 1.3 TiB 61.40 1.01 68 up .... omission .... TOTAL 63 TiB 38 TiB 3.8 GiB 23 TiB 16 TiB 25 TiB 60.96 MIN/MAX VAR: 0.68/1.30 STDDEV: 9.98 _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io