This [*]is the output of "ceph osd df".

Thanks a lot !
Massimo


[*]
[root@ceph-mon-01 ~]# ceph osd df
ID CLASS WEIGHT  REWEIGHT SIZE  USE   AVAIL %USE  VAR  PGS
30   hdd 5.45609  1.00000 5587G 1875G 3711G 33.57 0.65 140
31   hdd 5.45609  1.00000 5587G 3951G 1635G 70.72 1.38 144
32   hdd 5.45609  1.00000 5587G 3426G 2160G 61.33 1.19 127
33   hdd 5.45609  1.00000 5587G 3548G 2038G 63.51 1.24 167
34   hdd 5.45609  1.00000 5587G 1847G 3739G 33.06 0.64 121
35   hdd 5.45609  1.00000 5587G 2496G 3090G 44.68 0.87 161
36   hdd 5.45609  1.00000 5587G 3038G 2548G 54.38 1.06 153
37   hdd 5.45609  1.00000 5587G 2834G 2752G 50.73 0.99 122
38   hdd 5.45609  1.00000 5587G 2781G 2805G 49.79 0.97 124
39   hdd 5.45609  1.00000 5587G 3362G 2224G 60.18 1.17 141
40   hdd 5.45609  1.00000 5587G 2738G 2848G 49.02 0.95 139
41   hdd 5.45609  1.00000 5587G 2924G 2662G 52.35 1.02 129
42   hdd 5.45609  1.00000 5587G 2195G 3391G 39.29 0.77 116
43   hdd 5.45609  1.00000 5587G 2654G 2932G 47.51 0.93 132
44   hdd 5.45609  1.00000 5587G 3180G 2406G 56.93 1.11 125
45   hdd 5.45609  1.00000 5587G 2727G 2859G 48.82 0.95 152
46   hdd 5.45609  1.00000 5587G 2844G 2742G 50.91 0.99 153
47   hdd 5.45609  1.00000 5587G 2611G 2975G 46.74 0.91 127
48   hdd 5.45609  1.00000 5587G 3575G 2011G 63.99 1.25 139
49   hdd 5.45609  1.00000 5587G 1876G 3710G 33.59 0.65 121
10   hdd 5.45609  1.00000 5587G 2884G 2702G 51.64 1.01 128
11   hdd 5.45609  1.00000 5587G 3401G 2185G 60.89 1.19 130
12   hdd 5.45609  1.00000 5587G 4023G 1563G 72.01 1.40 153
13   hdd 5.45609  1.00000 5587G 1303G 4283G 23.34 0.45 131
14   hdd 5.45609  1.00000 5587G 2792G 2794G 49.97 0.97 135
15   hdd 5.45609  1.00000 5587G 1765G 3821G 31.61 0.62 123
16   hdd 5.45609  1.00000 5587G 3958G 1628G 70.86 1.38 152
17   hdd 5.45609  1.00000 5587G 4362G 1224G 78.09 1.52 139
18   hdd 5.45609  1.00000 5587G 2766G 2820G 49.51 0.96 144
19   hdd 5.45609  1.00000 5587G 3427G 2159G 61.34 1.19 131
20   hdd 5.45609  1.00000 5587G 3226G 2360G 57.75 1.12 162
21   hdd 5.45609  1.00000 5587G 2247G 3339G 40.22 0.78 146
22   hdd 5.45609  1.00000 5587G 2128G 3458G 38.10 0.74 124
23   hdd 5.45609  1.00000 5587G 2749G 2837G 49.21 0.96 133
24   hdd 5.45609  1.00000 5587G 3979G 1607G 71.24 1.39 148
25   hdd 5.45609  1.00000 5587G 2179G 3407G 39.02 0.76 121
26   hdd 5.45609  1.00000 5587G 3860G 1726G 69.09 1.35 151
27   hdd 5.45609  1.00000 5587G 2161G 3425G 38.68 0.75 137
28   hdd 5.45609  1.00000 5587G 3898G 1688G 69.78 1.36 141
29   hdd 5.45609  1.00000 5587G 2355G 3231G 42.15 0.82 121
 0   hdd 5.45609  1.00000 5587G 3294G 2292G 58.97 1.15 127
 1   hdd 5.45609  1.00000 5587G 2515G 3071G 45.02 0.88 132
 2   hdd 5.45609  1.00000 5587G 3300G 2286G 59.07 1.15 144
 3   hdd 5.45609  1.00000 5587G 2943G 2643G 52.68 1.03 151
 4   hdd 5.45609  1.00000 5587G 2641G 2945G 47.29 0.92 114
 5   hdd 5.45609  1.00000 5587G 2786G 2801G 49.87 0.97 131
 6   hdd 5.45609  1.00000 5587G 2564G 3022G 45.90 0.89 121
 7   hdd 5.45609  1.00000 5587G 1923G 3663G 34.43 0.67 143
 8   hdd 5.45609  1.00000 5587G 2625G 2961G 46.99 0.91 130
 9   hdd 5.45609  1.00000 5587G 2921G 2665G 52.30 1.02 140
                    TOTAL  272T  140T  132T 51.36
MIN/MAX VAR: 0.45/1.52  STDDEV: 12.09


On Mon, Jan 14, 2019 at 3:22 PM Wido den Hollander <w...@42on.com> wrote:

>
>
> On 1/14/19 3:18 PM, Massimo Sgaravatto wrote:
> > Thanks for the prompt reply
> >
> > Indeed I have different racks with different weights.
> > Below the ceph osd tree" output
> >
>
> Can you also show the output of 'ceph osd df' ?
>
> The amount of PGs might be on the low side which also causes this
> imbalance.
>
> If you do not have enough PGs CRUSH can't properly distribute either.
>
> Wido
>
> > [root@ceph-mon-01 ~]# ceph osd tree
> > ID CLASS WEIGHT    TYPE NAME                 STATUS REWEIGHT PRI-AFF
> > -1       272.80426 root default
> > -7       109.12170     rack Rack11-PianoAlto
> > -8        54.56085         host ceph-osd-04
> > 30   hdd   5.45609             osd.30            up  1.00000 1.00000
> > 31   hdd   5.45609             osd.31            up  1.00000 1.00000
> > 32   hdd   5.45609             osd.32            up  1.00000 1.00000
> > 33   hdd   5.45609             osd.33            up  1.00000 1.00000
> > 34   hdd   5.45609             osd.34            up  1.00000 1.00000
> > 35   hdd   5.45609             osd.35            up  1.00000 1.00000
> > 36   hdd   5.45609             osd.36            up  1.00000 1.00000
> > 37   hdd   5.45609             osd.37            up  1.00000 1.00000
> > 38   hdd   5.45609             osd.38            up  1.00000 1.00000
> > 39   hdd   5.45609             osd.39            up  1.00000 1.00000
> > -9        54.56085         host ceph-osd-05
> > 40   hdd   5.45609             osd.40            up  1.00000 1.00000
> > 41   hdd   5.45609             osd.41            up  1.00000 1.00000
> > 42   hdd   5.45609             osd.42            up  1.00000 1.00000
> > 43   hdd   5.45609             osd.43            up  1.00000 1.00000
> > 44   hdd   5.45609             osd.44            up  1.00000 1.00000
> > 45   hdd   5.45609             osd.45            up  1.00000 1.00000
> > 46   hdd   5.45609             osd.46            up  1.00000 1.00000
> > 47   hdd   5.45609             osd.47            up  1.00000 1.00000
> > 48   hdd   5.45609             osd.48            up  1.00000 1.00000
> > 49   hdd   5.45609             osd.49            up  1.00000 1.00000
> > -6       109.12170     rack Rack15-PianoAlto
> > -3        54.56085         host ceph-osd-02
> > 10   hdd   5.45609             osd.10            up  1.00000 1.00000
> > 11   hdd   5.45609             osd.11            up  1.00000 1.00000
> > 12   hdd   5.45609             osd.12            up  1.00000 1.00000
> > 13   hdd   5.45609             osd.13            up  1.00000 1.00000
> > 14   hdd   5.45609             osd.14            up  1.00000 1.00000
> > 15   hdd   5.45609             osd.15            up  1.00000 1.00000
> > 16   hdd   5.45609             osd.16            up  1.00000 1.00000
> > 17   hdd   5.45609             osd.17            up  1.00000 1.00000
> > 18   hdd   5.45609             osd.18            up  1.00000 1.00000
> > 19   hdd   5.45609             osd.19            up  1.00000 1.00000
> > -4        54.56085         host ceph-osd-03
> > 20   hdd   5.45609             osd.20            up  1.00000 1.00000
> > 21   hdd   5.45609             osd.21            up  1.00000 1.00000
> > 22   hdd   5.45609             osd.22            up  1.00000 1.00000
> > 23   hdd   5.45609             osd.23            up  1.00000 1.00000
> > 24   hdd   5.45609             osd.24            up  1.00000 1.00000
> > 25   hdd   5.45609             osd.25            up  1.00000 1.00000
> > 26   hdd   5.45609             osd.26            up  1.00000 1.00000
> > 27   hdd   5.45609             osd.27            up  1.00000 1.00000
> > 28   hdd   5.45609             osd.28            up  1.00000 1.00000
> > 29   hdd   5.45609             osd.29            up  1.00000 1.00000
> > -5        54.56085     rack Rack17-PianoAlto
> > -2        54.56085         host ceph-osd-01
> >  0   hdd   5.45609             osd.0             up  1.00000 1.00000
> >  1   hdd   5.45609             osd.1             up  1.00000 1.00000
> >  2   hdd   5.45609             osd.2             up  1.00000 1.00000
> >  3   hdd   5.45609             osd.3             up  1.00000 1.00000
> >  4   hdd   5.45609             osd.4             up  1.00000 1.00000
> >  5   hdd   5.45609             osd.5             up  1.00000 1.00000
> >  6   hdd   5.45609             osd.6             up  1.00000 1.00000
> >  7   hdd   5.45609             osd.7             up  1.00000 1.00000
> >  8   hdd   5.45609             osd.8             up  1.00000 1.00000
> >  9   hdd   5.45609             osd.9             up  1.00000 1.00000
> > [root@ceph-mon-01 ~]#
> >
> > On Mon, Jan 14, 2019 at 3:13 PM Dan van der Ster <d...@vanderster.com
> > <mailto:d...@vanderster.com>> wrote:
> >
> >     On Mon, Jan 14, 2019 at 3:06 PM Massimo Sgaravatto
> >     <massimo.sgarava...@gmail.com <mailto:massimo.sgarava...@gmail.com>>
> >     wrote:
> >     >
> >     > I have a ceph luminous cluster running on CentOS7 nodes.
> >     > This cluster has 50 OSDs, all with the same size and all with the
> >     same weight.
> >     >
> >     > Since I noticed that there was a quite "unfair" usage of OSD nodes
> >     (some used at 30 %, some used at 70 %) I tried to activate the
> balancer.
> >     >
> >     > But the balancer doesn't start I guess because of this problem:
> >     >
> >     > [root@ceph-mon-01 ~]# ceph osd crush weight-set create-compat
> >     > Error EPERM: crush map contains one or more bucket(s) that are not
> >     straw2
> >     >
> >     >
> >     > So I issued the command to convert from straw to straw2 (all the
> >     clients are running luminous):
> >     >
> >     >
> >     > [root@ceph-mon-01 ~]# ceph osd crush
> set-all-straw-buckets-to-straw2
> >     > Error EINVAL: new crush map requires client version hammer but
> >     require_min_compat_client is firefly
> >     > [root@ceph-mon-01 ~]# ceph osd set-require-min-compat-client jewel
> >     > set require_min_compat_client to jewel
> >     > [root@ceph-mon-01 ~]# ceph osd crush
> set-all-straw-buckets-to-straw2
> >     > [root@ceph-mon-01 ~]#
> >     >
> >     >
> >     > After having issued the command, the cluster went in WARNING state
> >     because ~ 12 % objects were misplaced.
> >     >
> >     > Is this normal ?
> >     > I read somewhere that the migration from straw to straw2 should
> >     trigger a data migration only if the OSDs have different sizes,
> >     which is not my case.
> >
> >     The relevant sizes to compare are the crush buckets across which you
> >     are replicating.
> >     Are you replicating host-wise or rack-wise?
> >     Do you have hosts/racks with a different crush weight (e.g. different
> >     crush size).
> >     Maybe share your `ceph osd tree`.
> >
> >     Cheers, dan
> >
> >
> >
> >     >
> >     >
> >     > The cluster is still recovering, but what is worrying me is that
> >     it looks like that data are being moved to the most used OSDs and
> >     the MAX_AVAIL value is decreasing quite quickly.
> >     >
> >     > I hope that the recovery can finish without causing problems: then
> >     I will immediately activate the balancer.
> >     >
> >     > But, if some OSDs are getting too full, is it safe to decrease
> >     their weights  while the cluster is still being recovered ?
> >     >
> >     > Thanks a lot for your help
> >     > Of course I can provide other info, if needed
> >     >
> >     >
> >     > Cheers, Massimo
> >     >
> >     > _______________________________________________
> >     > ceph-users mailing list
> >     > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> >     > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to