Hi Félix,

Could you please send me the output of the "ceph report" command (privately, 
the output is likely too big for the list) ? I suspect what you're seeing is 
because the smaller disks have more PGs than they should for the 
default.rgw.buckets.data pool. With the output of "ceph report" and using 
python-crush[1] we'll be able to verify that.

Cheers

[1] http://crush.readthedocs.io/

On 06/05/2017 01:54 PM, Félix Barbeira wrote:
> Hi,
> 
> We have a small cluster for radosgw use only. It has three nodes, witch 3 
> osds each. Each node has different disk sizes:
> 
> node01 : 3x8TB
> node02 : 3x2TB
> node03 : 3x3TB
> 
> I thought that the weight handle the amount of data that every osd receive. 
> In this case for example the node with the 8TB disks should receive more than 
> the rest, right? All of them receive the same amount of data and the smaller 
> disk (2TB) reaches 100% before the bigger ones. Am I doing something wrong?
> 
> The cluster is jewel LTS 10.2.7.
> 
> # ceph osd df
> ID WEIGHT  REWEIGHT SIZE   USE   AVAIL  %USE  VAR  PGS
>  0 7.27060  1.00000  7445G 1012G  6432G 13.60 0.57 133
>  3 7.27060  1.00000  7445G 1081G  6363G 14.52 0.61 163
>  4 7.27060  1.00000  7445G  787G  6657G 10.58 0.44 120
>  1 1.81310  1.00000  1856G 1047G   809G 56.41 2.37 143
>  5 1.81310  1.00000  1856G  956G   899G 51.53 2.16 143
>  6 1.81310  1.00000  1856G  877G   979G 47.24 1.98 130
>  2 2.72229  1.00000  2787G 1010G  1776G 36.25 1.52 140
>  7 2.72229  1.00000  2787G  831G  1955G 29.83 1.25 130
>  8 2.72229  1.00000  2787G 1038G  1748G 37.27 1.56 146
>               TOTAL 36267G 8643G 27624G 23.83
> MIN/MAX VAR: 0.44/2.37  STDDEV: 18.60
> #
> 
> # ceph osd tree
> ID WEIGHT   TYPE NAME      UP/DOWN REWEIGHT PRIMARY-AFFINITY
> -1 35.41795 root default
> -2 21.81180     host node01
>  0  7.27060         osd.0       up  1.00000          1.00000
>  3  7.27060         osd.3       up  1.00000          1.00000
>  4  7.27060         osd.4       up  1.00000          1.00000
> -3  5.43929     host node02
>  1  1.81310         osd.1       up  1.00000          1.00000
>  5  1.81310         osd.5       up  1.00000          1.00000
>  6  1.81310         osd.6       up  1.00000          1.00000
> -4  8.16687     host node03
>  2  2.72229         osd.2       up  1.00000          1.00000
>  7  2.72229         osd.7       up  1.00000          1.00000
>  8  2.72229         osd.8       up  1.00000          1.00000
> #
> 
> # ceph -s
>     cluster 49ba9695-7199-4c21-9199-ac321e60065e
>      health HEALTH_OK
>      monmap e1: 3 mons at 
> {ceph-mon01=[x:x:x:x:x:x:x:x]:6789/0,ceph-mon02=[x:x:x:x:x:x:x:x]:6789/0,ceph-mon03=[x:x:x:x:x:x:x:x]:6789/0}
>             election epoch 48, quorum 0,1,2 ceph-mon01,ceph-mon03,ceph-mon02
>      osdmap e265: 9 osds: 9 up, 9 in
>             flags sortbitwise,require_jewel_osds
>       pgmap v95701: 416 pgs, 11 pools, 2879 GB data, 729 kobjects
>             8643 GB used, 27624 GB / 36267 GB avail
>                  416 active+clean
> #
> 
> # ceph osd pool ls
> .rgw.root
> default.rgw.control
> default.rgw.data.root
> default.rgw.gc
> default.rgw.log
> default.rgw.users.uid
> default.rgw.users.keys
> default.rgw.buckets.index
> default.rgw.buckets.non-ec
> default.rgw.buckets.data
> default.rgw.users.email
> #
> 
> # ceph df
> GLOBAL:
>     SIZE       AVAIL      RAW USED     %RAW USED
>     36267G     27624G        8643G         23.83
> POOLS:
>     NAME                           ID     USED      %USED     MAX AVAIL     
> OBJECTS
>     .rgw.root                      1       1588         0         5269G       
>     4
>     default.rgw.control            2          0         0         5269G       
>     8
>     default.rgw.data.root          3       8761         0         5269G       
>    28
>     default.rgw.gc                 4          0         0         5269G       
>    32
>     default.rgw.log                5          0         0         5269G       
>   127
>     default.rgw.users.uid          6       4887         0         5269G       
>    28
>     default.rgw.users.keys         7        144         0         5269G       
>    16
>     default.rgw.buckets.index      9          0         0         5269G       
>    14
>     default.rgw.buckets.non-ec     10         0         0         5269G       
>     3
>     default.rgw.buckets.data       11     2879G     35.34         5269G      
> 746848
>     default.rgw.users.email        12        13         0         5269G       
>     1
> #
> 
> -- 
> Félix Barbeira.
> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 

-- 
Loïc Dachary, Artisan Logiciel Libre
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to