While you're watching things, if an OSD is getting too close for
comfort to the full ratio, you can temporarily increase it, e.g.
    ceph osd set-full-ratio 0.96

But don't set that too high -- you can really break an OSD if it gets
100% full (and then can't delete objects or whatever...)

-- dan

On Tue, Mar 23, 2021 at 7:17 PM Boris Behrens <b...@kervyn.de> wrote:
>
> Ok, then I will try to reweight the most filled OSDs to .95 and see if this 
> helps.
>
> Am Di., 23. März 2021 um 19:13 Uhr schrieb Dan van der Ster 
> <d...@vanderster.com>:
>>
>> Data goes to *all* PGs uniformly.
>> Max_avail is limited by the available space on the most full OSD --
>> you should pay close attention to those and make sure they are moving
>> in the right direction (decreasing!)
>>
>> Another point -- IMHO you should aim to get all PGs active+clean
>> before you add yet another batch of new disks. While there are PGs
>> backfilling, your osdmaps are accumulating on the mons and osds --
>> this itself will start to use a lot of space, and active+clean is the
>> only way to trim the old maps.
>>
>> -- dan
>>
>> On Tue, Mar 23, 2021 at 7:05 PM Boris Behrens <b...@kervyn.de> wrote:
>> >
>> > So,
>> > doing nothing and wait for the ceph to recover?
>> >
>> > In theory there should be enough disk space (more disks arriving 
>> > tomorrow), but I fear that there might be an issue, when the backups get 
>> > exported over night to this s3. Currently the max_avail lingers around 
>> > 13TB and I hope, that the data will go to other PGs than the ones that are 
>> > currently on filled OSDs.
>> >
>> >
>> >
>> > Am Di., 23. März 2021 um 18:58 Uhr schrieb Dan van der Ster 
>> > <d...@vanderster.com>:
>> >>
>> >> Hi,
>> >>
>> >> backfill_toofull is not a bad thing when the cluster is really full
>> >> like yours. You should expect some of the most full OSDs to eventually
>> >> start decreasing in usage, as the PGs are moved to the new OSDs. Those
>> >> backfill_toofull states should then resolve themselves as the OSD
>> >> usage flattens out.
>> >> Keep an eye on the usage of the backfill_full and nearfull OSDs though
>> >> -- if they do eventually go above the full_ratio (95% by default),
>> >> then writes to those OSDs would stop.
>> >>
>> >> But if on the other hand you're suffering from lots of slow ops or
>> >> anything else visible to your users, then you could try to take some
>> >> actions to slow down the rebalancing. Just let us know if that's the
>> >> case and we can see about changing osd_max_backfills, some weights or
>> >> maybe using the upmap-remapped tool.
>> >>
>> >> -- Dan
>> >>
>> >> On Tue, Mar 23, 2021 at 6:07 PM Boris Behrens <b...@kervyn.de> wrote:
>> >> >
>> >> > Ok, I should have listened to you :)
>> >> >
>> >> > In the last week we added more storage but the issue got worse instead.
>> >> > Today I realized that the PGs were up to 90GB (bytes column in ceph pg 
>> >> > ls said 95705749636), and the balance kept mentioning the 2048 PGs for 
>> >> > this pool. We were at 72% utilization (ceph osd df tree, first line) 
>> >> > for our cluster and I increased the PGs to 2048.
>> >> >
>> >> > Now I am in a world of trouble.
>> >> > The space in the cluster went down, I am at 45% misplaced objects, and 
>> >> > we already added 20x4TB disks just to not go down completly.
>> >> >
>> >> > The utilization is still going up and the overall free space in the 
>> >> > cluster seems to go down. This is what my ceph status looks like and 
>> >> > now I really need help to get that thing back to normal:
>> >> > [root@s3db1 ~]# ceph status
>> >> >   cluster:
>> >> >     id:     dca79fff-ffd0-58f4-1cff-82a2feea05f4
>> >> >     health: HEALTH_WARN
>> >> >             4 backfillfull osd(s)
>> >> >             17 nearfull osd(s)
>> >> >             37 pool(s) backfillfull
>> >> >             13 large omap objects
>> >> >             Low space hindering backfill (add storage if this doesn't 
>> >> > resolve itself): 570 pgs backfill_toofull
>> >> >
>> >> >   services:
>> >> >     mon: 3 daemons, quorum ceph-s3-mon1,ceph-s3-mon2,ceph-s3-mon3 (age 
>> >> > 44m)
>> >> >     mgr: ceph-mgr2(active, since 15m), standbys: ceph-mgr3, ceph-mgr1
>> >> >     mds:  3 up:standby
>> >> >     osd: 110 osds: 110 up (since 28m), 110 in (since 28m); 1535 
>> >> > remapped pgs
>> >> >     rgw: 3 daemons active (eu-central-1, eu-msg-1, eu-secure-1)
>> >> >
>> >> >   task status:
>> >> >
>> >> >   data:
>> >> >     pools:   37 pools, 4032 pgs
>> >> >     objects: 116.23M objects, 182 TiB
>> >> >     usage:   589 TiB used, 206 TiB / 795 TiB avail
>> >> >     pgs:     160918554/348689415 objects misplaced (46.150%)
>> >> >              2497 active+clean
>> >> >              779  active+remapped+backfill_wait
>> >> >              538  active+remapped+backfill_wait+backfill_toofull
>> >> >              186  active+remapped+backfilling
>> >> >              32   active+remapped+backfill_toofull
>> >> >
>> >> >   io:
>> >> >     client:   27 MiB/s rd, 69 MiB/s wr, 497 op/s rd, 153 op/s wr
>> >> >     recovery: 1.5 GiB/s, 922 objects/s
>> >> >
>> >> > Am Di., 16. März 2021 um 09:34 Uhr schrieb Boris Behrens 
>> >> > <b...@kervyn.de>:
>> >> >>
>> >> >> Hi Dan,
>> >> >>
>> >> >> my EC profile look very "default" to me.
>> >> >> [root@s3db1 ~]# ceph osd erasure-code-profile ls
>> >> >> default
>> >> >> [root@s3db1 ~]# ceph osd erasure-code-profile get default
>> >> >> k=2
>> >> >> m=1
>> >> >> plugin=jerasure
>> >> >> technique=reed_sol_van
>> >> >>
>> >> >> I don't understand the ouput, but the balancing get worse over night:
>> >> >>
>> >> >> [root@s3db1 ~]# ceph-scripts/tools/ceph-pool-pg-distribution 11
>> >> >> Searching for PGs in pools: ['11']
>> >> >> Summary: 1024 PGs on 84 osds
>> >> >>
>> >> >> Num OSDs with X PGs:
>> >> >> 15: 8
>> >> >> 16: 7
>> >> >> 17: 6
>> >> >> 18: 10
>> >> >> 19: 1
>> >> >> 32: 10
>> >> >> 33: 4
>> >> >> 34: 6
>> >> >> 35: 8
>> >> >> 65: 5
>> >> >> 66: 5
>> >> >> 67: 4
>> >> >> 68: 10
>> >> >> [root@s3db1 ~]# ceph-scripts/tools/ceph-pg-histogram --normalize 
>> >> >> --pool=11
>> >> >> # NumSamples = 84; Min = 4.12; Max = 5.09
>> >> >> # Mean = 4.553355; Variance = 0.052415; SD = 0.228942; Median 4.561608
>> >> >> # each ∎ represents a count of 1
>> >> >>     4.1244 -     4.2205 [     8]: ∎∎∎∎∎∎∎∎
>> >> >>     4.2205 -     4.3166 [     6]: ∎∎∎∎∎∎
>> >> >>     4.3166 -     4.4127 [    11]: ∎∎∎∎∎∎∎∎∎∎∎
>> >> >>     4.4127 -     4.5087 [    10]: ∎∎∎∎∎∎∎∎∎∎
>> >> >>     4.5087 -     4.6048 [    11]: ∎∎∎∎∎∎∎∎∎∎∎
>> >> >>     4.6048 -     4.7009 [    19]: ∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎
>> >> >>     4.7009 -     4.7970 [     6]: ∎∎∎∎∎∎
>> >> >>     4.7970 -     4.8931 [     8]: ∎∎∎∎∎∎∎∎
>> >> >>     4.8931 -     4.9892 [     4]: ∎∎∎∎
>> >> >>     4.9892 -     5.0852 [     1]: ∎
>> >> >> [root@s3db1 ~]# ceph osd df tree | sort -nk 17 | tail
>> >> >>  14   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 724 GiB   19 GiB     0 B 
>> >> >> 724 GiB 80.56 1.07  56     up         osd.14
>> >> >>  19   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB 2.9 TiB  466 MiB 7.9 GiB 
>> >> >> 708 GiB 81.25 1.08  53     up         osd.19
>> >> >>   4   hdd   3.63689  1.00000 3.6 TiB 3.0 TiB 698 GiB  703 MiB     0 B 
>> >> >> 698 GiB 81.27 1.08  48     up         osd.4
>> >> >>  24   hdd   3.63689  1.00000 3.6 TiB 3.0 TiB 695 GiB  640 MiB     0 B 
>> >> >> 695 GiB 81.34 1.08  46     up         osd.24
>> >> >>  75   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB 2.9 TiB  440 MiB 8.1 GiB 
>> >> >> 704 GiB 81.35 1.08  48     up         osd.75
>> >> >>  71   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB 3.0 TiB  7.5 MiB 8.0 GiB 
>> >> >> 663 GiB 82.44 1.09  47     up         osd.71
>> >> >>  76   hdd   3.68750  1.00000 3.7 TiB 3.1 TiB 3.0 TiB  251 MiB 9.0 GiB 
>> >> >> 617 GiB 83.65 1.11  50     up         osd.76
>> >> >>  33   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB 3.0 TiB  399 MiB 8.1 GiB 
>> >> >> 618 GiB 83.85 1.11  55     up         osd.33
>> >> >>  35   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB 3.0 TiB  317 MiB 8.8 GiB 
>> >> >> 617 GiB 83.87 1.11  50     up         osd.35
>> >> >>  34   hdd   3.73630  1.00000 3.7 TiB 3.2 TiB 3.1 TiB  451 MiB 8.7 GiB 
>> >> >> 545 GiB 85.75 1.14  54     up         osd.34
>> >> >>
>> >> >> Am Mo., 15. März 2021 um 17:23 Uhr schrieb Dan van der Ster 
>> >> >> <d...@vanderster.com>:
>> >> >>>
>> >> >>> Hi,
>> >> >>>
>> >> >>> How wide are your EC profiles? If they are really wide, you might be
>> >> >>> reaching the limits of what is physically possible. Also, I'm not sure
>> >> >>> that upmap in 14.2.11 is very smart about *improving* existing upmap
>> >> >>> rules for a given PG, in the case that a PG already has an upmap-items
>> >> >>> entry but it would help the distribution to add more mapping pairs to
>> >> >>> that entry. What this means, is that it might sometimes be useful to
>> >> >>> randomly remove some upmap entries and see if the balancer does a
>> >> >>> better job when it replaces them.
>> >> >>>
>> >> >>> But before you do that, I re-remembered that looking at the total PG
>> >> >>> numbers is not useful -- you need to check the PGs per OSD for the
>> >> >>> eu-central-1.rgw.buckets.data pool only.
>> >> >>>
>> >> >>> We have a couple tools that can help with this:
>> >> >>>
>> >> >>> 1. To see the PGs per OSD for a given pool:
>> >> >>>     
>> >> >>> https://github.com/cernceph/ceph-scripts/blob/master/tools/ceph-pool-pg-distribution
>> >> >>>
>> >> >>>     E.g.: ./ceph-pool-pg-distribution 11  # to see the distribution of
>> >> >>> your eu-central-1.rgw.buckets.data pool.
>> >> >>>
>> >> >>>     The output looks like this on my well balanced clusters:
>> >> >>>
>> >> >>> # ceph-scripts/tools/ceph-pool-pg-distribution 15
>> >> >>> Searching for PGs in pools: ['15']
>> >> >>> Summary: 256 pgs on 56 osds
>> >> >>>
>> >> >>> Num OSDs with X PGs:
>> >> >>>  13: 16
>> >> >>>  14: 40
>> >> >>>
>> >> >>>     You should expect a trimodal for your cluster.
>> >> >>>
>> >> >>> 2. You can also use another script from that repo to see the PGs per
>> >> >>> OSD normalized to crush weight:
>> >> >>>     ceph-scripts/tools/ceph-pg-histogram --normalize --pool=15
>> >> >>>
>> >> >>>    This might explain what is going wrong.
>> >> >>>
>> >> >>> Cheers, Dan
>> >> >>>
>> >> >>>
>> >> >>> On Mon, Mar 15, 2021 at 3:04 PM Boris Behrens <b...@kervyn.de> wrote:
>> >> >>> >
>> >> >>> > Absolutly:
>> >> >>> > [root@s3db1 ~]# ceph osd df tree
>> >> >>> > ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE DATA     OMAP     META 
>> >> >>> >    AVAIL    %USE  VAR  PGS STATUS TYPE NAME
>> >> >>> >  -1       673.54224        - 674 TiB 496 TiB  468 TiB   97 GiB 1.2 
>> >> >>> > TiB  177 TiB 73.67 1.00   -        root default
>> >> >>> >  -2        58.30331        -  58 TiB  42 TiB   38 TiB  9.2 GiB  99 
>> >> >>> > GiB   16 TiB 72.88 0.99   -            host s3db1
>> >> >>> >  23   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  714 MiB  25 
>> >> >>> > GiB  3.7 TiB 74.87 1.02 194     up         osd.23
>> >> >>> >  69   hdd  14.55269  1.00000  15 TiB  11 TiB   11 TiB  1.6 GiB  40 
>> >> >>> > GiB  3.4 TiB 76.32 1.04 199     up         osd.69
>> >> >>> >  73   hdd  14.55269  1.00000  15 TiB  11 TiB   11 TiB  1.3 GiB  34 
>> >> >>> > GiB  3.8 TiB 74.15 1.01 203     up         osd.73
>> >> >>> >  79   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.3 TiB  1.8 GiB     
>> >> >>> > 0 B  1.3 TiB 65.44 0.89  47     up         osd.79
>> >> >>> >  80   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.3 TiB  2.2 GiB     
>> >> >>> > 0 B  1.3 TiB 65.34 0.89  48     up         osd.80
>> >> >>> >  81   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.3 TiB  1.1 GiB     
>> >> >>> > 0 B  1.3 TiB 65.38 0.89  47     up         osd.81
>> >> >>> >  82   hdd   3.63689  1.00000 3.6 TiB 2.5 TiB  1.1 TiB  619 MiB     
>> >> >>> > 0 B  1.1 TiB 68.46 0.93  41     up         osd.82
>> >> >>> > -11        50.94173        -  51 TiB  37 TiB   37 TiB  3.5 GiB  98 
>> >> >>> > GiB   14 TiB 71.90 0.98   -            host s3db10
>> >> >>> >  63   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB  5.3 TiB  647 MiB  14 
>> >> >>> > GiB  2.0 TiB 72.72 0.99  94     up         osd.63
>> >> >>> >  64   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB  5.2 TiB  668 MiB  14 
>> >> >>> > GiB  2.0 TiB 72.23 0.98  93     up         osd.64
>> >> >>> >  65   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2 TiB  227 MiB  14 
>> >> >>> > GiB  2.1 TiB 71.16 0.97 100     up         osd.65
>> >> >>> >  66   hdd   7.27739  1.00000 7.3 TiB 5.4 TiB  5.4 TiB  313 MiB  14 
>> >> >>> > GiB  1.9 TiB 74.25 1.01  92     up         osd.66
>> >> >>> >  67   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB  5.1 TiB  584 MiB  14 
>> >> >>> > GiB  2.1 TiB 70.63 0.96  96     up         osd.67
>> >> >>> >  68   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2 TiB  720 MiB  14 
>> >> >>> > GiB  2.1 TiB 71.72 0.97 101     up         osd.68
>> >> >>> >  70   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB  5.1 TiB  425 MiB  14 
>> >> >>> > GiB  2.1 TiB 70.59 0.96  97     up         osd.70
>> >> >>> > -12        50.99052        -  51 TiB  38 TiB   37 TiB  2.1 GiB  97 
>> >> >>> > GiB   13 TiB 73.77 1.00   -            host s3db11
>> >> >>> >  46   hdd   7.27739  1.00000 7.3 TiB 5.6 TiB  5.6 TiB  229 MiB  14 
>> >> >>> > GiB  1.7 TiB 77.05 1.05  97     up         osd.46
>> >> >>> >  47   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB  5.1 TiB  159 MiB  13 
>> >> >>> > GiB  2.2 TiB 70.00 0.95  89     up         osd.47
>> >> >>> >  48   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2 TiB  279 MiB  14 
>> >> >>> > GiB  2.1 TiB 71.82 0.97  98     up         osd.48
>> >> >>> >  49   hdd   7.27739  1.00000 7.3 TiB 5.5 TiB  5.4 TiB  276 MiB  14 
>> >> >>> > GiB  1.8 TiB 74.90 1.02  95     up         osd.49
>> >> >>> >  50   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2 TiB  336 MiB  14 
>> >> >>> > GiB  2.0 TiB 72.13 0.98  93     up         osd.50
>> >> >>> >  51   hdd   7.27739  1.00000 7.3 TiB 5.7 TiB  5.6 TiB  728 MiB  15 
>> >> >>> > GiB  1.6 TiB 77.76 1.06  98     up         osd.51
>> >> >>> >  72   hdd   7.32619  1.00000 7.3 TiB 5.3 TiB  5.3 TiB  147 MiB  13 
>> >> >>> > GiB  2.0 TiB 72.75 0.99  95     up         osd.72
>> >> >>> > -37        58.55478        -  59 TiB  44 TiB   44 TiB  4.4 GiB 122 
>> >> >>> > GiB   15 TiB 75.20 1.02   -            host s3db12
>> >> >>> >  19   hdd   3.68750  1.00000 3.7 TiB 2.9 TiB  2.9 TiB  454 MiB 8.2 
>> >> >>> > GiB  780 GiB 79.35 1.08  53     up         osd.19
>> >> >>> >  71   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB  2.9 TiB  7.1 MiB 8.0 
>> >> >>> > GiB  734 GiB 80.56 1.09  47     up         osd.71
>> >> >>> >  75   hdd   3.68750  1.00000 3.7 TiB 2.9 TiB  2.9 TiB  439 MiB 8.2 
>> >> >>> > GiB  777 GiB 79.43 1.08  48     up         osd.75
>> >> >>> >  76   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB  3.0 TiB  241 MiB 8.9 
>> >> >>> > GiB  688 GiB 81.77 1.11  50     up         osd.76
>> >> >>> >  77   hdd  14.60159  1.00000  15 TiB  11 TiB   11 TiB  880 MiB  30 
>> >> >>> > GiB  3.6 TiB 75.44 1.02 201     up         osd.77
>> >> >>> >  78   hdd  14.60159  1.00000  15 TiB  10 TiB   10 TiB 1015 MiB  28 
>> >> >>> > GiB  4.2 TiB 71.26 0.97 193     up         osd.78
>> >> >>> >  83   hdd  14.60159  1.00000  15 TiB  11 TiB   11 TiB  1.4 GiB  30 
>> >> >>> > GiB  3.8 TiB 73.76 1.00 203     up         osd.83
>> >> >>> >  -3        58.49872        -  58 TiB  42 TiB   36 TiB  8.2 GiB  89 
>> >> >>> > GiB   17 TiB 71.71 0.97   -            host s3db2
>> >> >>> >   1   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  3.2 GiB  37 
>> >> >>> > GiB  3.7 TiB 74.58 1.01 196     up         osd.1
>> >> >>> >   3   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3 TiB  566 MiB     
>> >> >>> > 0 B  1.3 TiB 64.11 0.87  50     up         osd.3
>> >> >>> >   4   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB  771 GiB  695 MiB     
>> >> >>> > 0 B  771 GiB 79.30 1.08  48     up         osd.4
>> >> >>> >   5   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.2 TiB  482 MiB     
>> >> >>> > 0 B  1.2 TiB 66.51 0.90  49     up         osd.5
>> >> >>> >   6   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3 TiB  1.8 GiB     
>> >> >>> > 0 B  1.3 TiB 64.00 0.87  42     up         osd.6
>> >> >>> >   7   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  639 MiB  26 
>> >> >>> > GiB  4.0 TiB 72.44 0.98 192     up         osd.7
>> >> >>> >  74   hdd  14.65039  1.00000  15 TiB  10 TiB   10 TiB  907 MiB  26 
>> >> >>> > GiB  4.2 TiB 71.32 0.97 193     up         osd.74
>> >> >>> >  -4        58.49872        -  58 TiB  43 TiB   36 TiB   34 GiB  85 
>> >> >>> > GiB   16 TiB 72.69 0.99   -            host s3db3
>> >> >>> >   2   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  980 MiB  26 
>> >> >>> > GiB  3.8 TiB 74.36 1.01 203     up         osd.2
>> >> >>> >   9   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  8.4 GiB  33 
>> >> >>> > GiB  3.9 TiB 73.51 1.00 186     up         osd.9
>> >> >>> >  10   hdd  14.65039  1.00000  15 TiB  10 TiB   10 TiB  650 MiB  26 
>> >> >>> > GiB  4.2 TiB 71.64 0.97 201     up         osd.10
>> >> >>> >  12   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3 TiB  754 MiB     
>> >> >>> > 0 B  1.3 TiB 64.17 0.87  44     up         osd.12
>> >> >>> >  13   hdd   3.63689  1.00000 3.6 TiB 2.8 TiB  813 GiB  2.4 GiB     
>> >> >>> > 0 B  813 GiB 78.17 1.06  58     up         osd.13
>> >> >>> >  14   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB  797 GiB   19 GiB     
>> >> >>> > 0 B  797 GiB 78.60 1.07  56     up         osd.14
>> >> >>> >  15   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3 TiB  2.2 GiB     
>> >> >>> > 0 B  1.3 TiB 63.96 0.87  41     up         osd.15
>> >> >>> >  -5        58.49872        -  58 TiB  43 TiB   36 TiB  6.7 GiB  97 
>> >> >>> > GiB   15 TiB 74.04 1.01   -            host s3db4
>> >> >>> >  11   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  940 MiB  26 
>> >> >>> > GiB  4.0 TiB 72.49 0.98 196     up         osd.11
>> >> >>> >  17   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB 1022 MiB  26 
>> >> >>> > GiB  3.6 TiB 75.23 1.02 204     up         osd.17
>> >> >>> >  18   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  945 MiB  45 
>> >> >>> > GiB  3.8 TiB 74.16 1.01 193     up         osd.18
>> >> >>> >  20   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1020 GiB  596 MiB     
>> >> >>> > 0 B 1020 GiB 72.62 0.99  57     up         osd.20
>> >> >>> >  21   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1023 GiB  1.9 GiB     
>> >> >>> > 0 B 1023 GiB 72.54 0.98  41     up         osd.21
>> >> >>> >  22   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1023 GiB  797 MiB     
>> >> >>> > 0 B 1023 GiB 72.54 0.98  53     up         osd.22
>> >> >>> >  24   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB  766 GiB  618 MiB     
>> >> >>> > 0 B  766 GiB 79.42 1.08  46     up         osd.24
>> >> >>> >  -6        58.89636        -  59 TiB  43 TiB   43 TiB  3.0 GiB 108 
>> >> >>> > GiB   16 TiB 73.40 1.00   -            host s3db5
>> >> >>> >   0   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.6 TiB   92 MiB 7.2 
>> >> >>> > GiB  1.1 TiB 71.16 0.97  45     up         osd.0
>> >> >>> >  25   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.6 TiB  2.4 MiB 7.3 
>> >> >>> > GiB  1.1 TiB 71.23 0.97  41     up         osd.25
>> >> >>> >  26   hdd   3.73630  1.00000 3.7 TiB 2.8 TiB  2.7 TiB  181 MiB 7.6 
>> >> >>> > GiB  935 GiB 75.57 1.03  45     up         osd.26
>> >> >>> >  27   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.6 TiB  5.1 MiB 7.0 
>> >> >>> > GiB  1.1 TiB 71.20 0.97  47     up         osd.27
>> >> >>> >  28   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  977 MiB  26 
>> >> >>> > GiB  3.8 TiB 73.85 1.00 197     up         osd.28
>> >> >>> >  29   hdd  14.65039  1.00000  15 TiB  11 TiB   10 TiB  872 MiB  26 
>> >> >>> > GiB  4.1 TiB 71.98 0.98 196     up         osd.29
>> >> >>> >  30   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  943 MiB  27 
>> >> >>> > GiB  3.6 TiB 75.51 1.03 202     up         osd.30
>> >> >>> >  -7        58.89636        -  59 TiB  44 TiB   43 TiB   13 GiB 122 
>> >> >>> > GiB   15 TiB 74.97 1.02   -            host s3db6
>> >> >>> >  32   hdd   3.73630  1.00000 3.7 TiB 2.8 TiB  2.7 TiB   27 MiB 7.6 
>> >> >>> > GiB  940 GiB 75.42 1.02  55     up         osd.32
>> >> >>> >  33   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB  3.0 TiB  376 MiB 8.2 
>> >> >>> > GiB  691 GiB 81.94 1.11  55     up         osd.33
>> >> >>> >  34   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB  3.0 TiB  450 MiB 8.5 
>> >> >>> > GiB  620 GiB 83.79 1.14  54     up         osd.34
>> >> >>> >  35   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB  3.0 TiB  316 MiB 8.4 
>> >> >>> > GiB  690 GiB 81.98 1.11  50     up         osd.35
>> >> >>> >  36   hdd  14.65039  1.00000  15 TiB  11 TiB   10 TiB  489 MiB  25 
>> >> >>> > GiB  4.1 TiB 71.69 0.97 208     up         osd.36
>> >> >>> >  37   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB   11 GiB  38 
>> >> >>> > GiB  4.0 TiB 72.41 0.98 195     up         osd.37
>> >> >>> >  38   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  1.1 GiB  26 
>> >> >>> > GiB  3.7 TiB 74.88 1.02 204     up         osd.38
>> >> >>> >  -8        58.89636        -  59 TiB  44 TiB   43 TiB  3.8 GiB 111 
>> >> >>> > GiB   15 TiB 74.16 1.01   -            host s3db7
>> >> >>> >  39   hdd   3.73630  1.00000 3.7 TiB 2.8 TiB  2.7 TiB   19 MiB 7.5 
>> >> >>> > GiB  936 GiB 75.54 1.03  39     up         osd.39
>> >> >>> >  40   hdd   3.73630  1.00000 3.7 TiB 2.6 TiB  2.5 TiB  144 MiB 7.1 
>> >> >>> > GiB  1.1 TiB 69.87 0.95  39     up         osd.40
>> >> >>> >  41   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.7 TiB  219 MiB 7.6 
>> >> >>> > GiB 1011 GiB 73.57 1.00  55     up         osd.41
>> >> >>> >  42   hdd   3.73630  1.00000 3.7 TiB 2.6 TiB  2.5 TiB  593 MiB 7.1 
>> >> >>> > GiB  1.1 TiB 70.02 0.95  47     up         osd.42
>> >> >>> >  43   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  500 MiB  27 
>> >> >>> > GiB  3.7 TiB 74.67 1.01 204     up         osd.43
>> >> >>> >  44   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  1.1 GiB  27 
>> >> >>> > GiB  3.7 TiB 74.62 1.01 193     up         osd.44
>> >> >>> >  45   hdd  14.65039  1.00000  15 TiB  11 TiB   11 TiB  1.2 GiB  29 
>> >> >>> > GiB  3.6 TiB 75.16 1.02 204     up         osd.45
>> >> >>> >  -9        51.28331        -  51 TiB  39 TiB   39 TiB  4.9 GiB 107 
>> >> >>> > GiB   12 TiB 76.50 1.04   -            host s3db8
>> >> >>> >   8   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.5 TiB  474 MiB  14 
>> >> >>> > GiB  1.7 TiB 76.37 1.04  98     up         osd.8
>> >> >>> >  16   hdd   7.32619  1.00000 7.3 TiB 5.7 TiB  5.7 TiB  783 MiB  15 
>> >> >>> > GiB  1.6 TiB 78.39 1.06 100     up         osd.16
>> >> >>> >  31   hdd   7.32619  1.00000 7.3 TiB 5.7 TiB  5.6 TiB  441 MiB  14 
>> >> >>> > GiB  1.6 TiB 77.70 1.05  91     up         osd.31
>> >> >>> >  52   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.5 TiB  939 MiB  14 
>> >> >>> > GiB  1.7 TiB 76.29 1.04 102     up         osd.52
>> >> >>> >  53   hdd   7.32619  1.00000 7.3 TiB 5.4 TiB  5.4 TiB  848 MiB  18 
>> >> >>> > GiB  1.9 TiB 74.30 1.01  98     up         osd.53
>> >> >>> >  54   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.6 TiB  1.0 GiB  16 
>> >> >>> > GiB  1.7 TiB 76.99 1.05 106     up         osd.54
>> >> >>> >  55   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB  5.5 TiB  460 MiB  15 
>> >> >>> > GiB  1.8 TiB 75.46 1.02 105     up         osd.55
>> >> >>> > -10        51.28331        -  51 TiB  37 TiB   37 TiB  3.8 GiB  96 
>> >> >>> > GiB   14 TiB 72.77 0.99   -            host s3db9
>> >> >>> >  56   hdd   7.32619  1.00000 7.3 TiB 5.2 TiB  5.2 TiB  846 MiB  13 
>> >> >>> > GiB  2.1 TiB 71.16 0.97 104     up         osd.56
>> >> >>> >  57   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.6 TiB  513 MiB  15 
>> >> >>> > GiB  1.7 TiB 76.53 1.04  96     up         osd.57
>> >> >>> >  58   hdd   7.32619  1.00000 7.3 TiB 5.2 TiB  5.2 TiB  604 MiB  13 
>> >> >>> > GiB  2.1 TiB 71.23 0.97  98     up         osd.58
>> >> >>> >  59   hdd   7.32619  1.00000 7.3 TiB 5.1 TiB  5.1 TiB  414 MiB  13 
>> >> >>> > GiB  2.2 TiB 70.03 0.95  88     up         osd.59
>> >> >>> >  60   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB  5.5 TiB  227 MiB  14 
>> >> >>> > GiB  1.8 TiB 75.54 1.03  97     up         osd.60
>> >> >>> >  61   hdd   7.32619  1.00000 7.3 TiB 5.1 TiB  5.1 TiB  456 MiB  13 
>> >> >>> > GiB  2.2 TiB 70.01 0.95  95     up         osd.61
>> >> >>> >  62   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB  5.4 TiB  843 MiB  14 
>> >> >>> > GiB  1.8 TiB 74.93 1.02 110     up         osd.62
>> >> >>> >                        TOTAL 674 TiB 496 TiB  468 TiB   97 GiB 1.2 
>> >> >>> > TiB  177 TiB 73.67
>> >> >>> > MIN/MAX VAR: 0.87/1.14  STDDEV: 4.22
>> >> >>> >
>> >> >>> > Am Mo., 15. März 2021 um 15:02 Uhr schrieb Dan van der Ster 
>> >> >>> > <d...@vanderster.com>:
>> >> >>> >>
>> >> >>> >> OK thanks. Indeed "prepared 0/10 changes" means it thinks things 
>> >> >>> >> are balanced.
>> >> >>> >> Could you again share the full ceph osd df tree?
>> >> >>> >>
>> >> >>> >> On Mon, Mar 15, 2021 at 2:54 PM Boris Behrens <b...@kervyn.de> 
>> >> >>> >> wrote:
>> >> >>> >> >
>> >> >>> >> > Hi Dan,
>> >> >>> >> >
>> >> >>> >> > I've set the autoscaler to warn, but it actually does not warn 
>> >> >>> >> > for now. So not touching it for now.
>> >> >>> >> >
>> >> >>> >> > this is what the log says in minute intervals:
>> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/active
>> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/sleep_interval
>> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/begin_time
>> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/end_time
>> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/begin_weekday
>> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/end_weekday
>> >> >>> >> > 2021-03-15 13:51:00.971 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/pool_ids
>> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] Optimize 
>> >> >>> >> > plan auto_2021-03-15_13:51:00
>> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/mode
>> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] Mode 
>> >> >>> >> > upmap, max misplaced 0.050000
>> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] do_upmap
>> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/upmap_max_iterations
>> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr get_config 
>> >> >>> >> > get_config key: mgr/balancer/upmap_max_deviation
>> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] pools 
>> >> >>> >> > ['eu-msg-1.rgw.data.root', 'eu-msg-1.rgw.buckets.non-ec', 
>> >> >>> >> > 'eu-central-1.rgw.users.keys', 'eu-central-1.rgw.gc', 
>> >> >>> >> > 'eu-central-1.rgw.buckets.data', 'eu-central-1.rgw.users.email', 
>> >> >>> >> > 'eu-msg-1.rgw.gc', 'eu-central-1.rgw.usage', 
>> >> >>> >> > 'eu-msg-1.rgw.users.keys', 'eu-central-1.rgw.buckets.index', 
>> >> >>> >> > 'rbd', 'eu-msg-1.rgw.log', 'whitespace-again-2021-03-10_2', 
>> >> >>> >> > 'eu-msg-1.rgw.buckets.index', 'eu-msg-1.rgw.meta', 
>> >> >>> >> > 'eu-central-1.rgw.log', 'default.rgw.gc', 
>> >> >>> >> > 'eu-central-1.rgw.buckets.non-ec', 'eu-msg-1.rgw.usage', 
>> >> >>> >> > 'whitespace-again-2021-03-10', 'fra-1.rgw.meta', 
>> >> >>> >> > 'eu-central-1.rgw.users.uid', 'eu-msg-1.rgw.users.email', 
>> >> >>> >> > 'fra-1.rgw.control', 'eu-msg-1.rgw.users.uid', 
>> >> >>> >> > 'eu-msg-1.rgw.control', '.rgw.root', 
>> >> >>> >> > 'eu-msg-1.rgw.buckets.data', 'default.rgw.control', 
>> >> >>> >> > 'fra-1.rgw.log', 'default.rgw.data.root', 
>> >> >>> >> > 'whitespace-again-2021-03-10_3', 'default.rgw.log', 
>> >> >>> >> > 'eu-central-1.rgw.meta', 'eu-central-1.rgw.data.root', 
>> >> >>> >> > 'default.rgw.users.uid', 'eu-central-1.rgw.control']
>> >> >>> >> > 2021-03-15 13:51:01.224 7f307d5fd700  4 mgr[balancer] prepared 
>> >> >>> >> > 0/10 changes
>> >> >>> >> >
>> >> >>> >> > Am Mo., 15. März 2021 um 14:15 Uhr schrieb Dan van der Ster 
>> >> >>> >> > <d...@vanderster.com>:
>> >> >>> >> >>
>> >> >>> >> >> I suggest to just disable the autoscaler until your balancing 
>> >> >>> >> >> is understood.
>> >> >>> >> >>
>> >> >>> >> >> What does your active mgr log say (with debug_mgr 4/5), grep 
>> >> >>> >> >> balancer
>> >> >>> >> >> /var/log/ceph/ceph-mgr.*.log
>> >> >>> >> >>
>> >> >>> >> >> -- Dan
>> >> >>> >> >>
>> >> >>> >> >> On Mon, Mar 15, 2021 at 1:47 PM Boris Behrens <b...@kervyn.de> 
>> >> >>> >> >> wrote:
>> >> >>> >> >> >
>> >> >>> >> >> > Hi,
>> >> >>> >> >> > this unfortunally did not solve my problem. I still have some 
>> >> >>> >> >> > OSDs that fill up to 85%
>> >> >>> >> >> >
>> >> >>> >> >> > According to the logging, the autoscaler might want to add 
>> >> >>> >> >> > more PGs to one Bucken and reduce almost all other buckets to 
>> >> >>> >> >> > 32.
>> >> >>> >> >> > 2021-03-15 12:19:58.825 7f307f601700  4 mgr[pg_autoscaler] 
>> >> >>> >> >> > Pool 'eu-central-1.rgw.buckets.data' root_id -1 using 
>> >> >>> >> >> > 0.705080476146 of space, bias 1.0, pg target 1974.22533321 
>> >> >>> >> >> > quantized to 2048 (current 1024)
>> >> >>> >> >> >
>> >> >>> >> >> > Why the balancing does not happen is still nebulous to me.
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> > Am Sa., 13. März 2021 um 16:37 Uhr schrieb Dan van der Ster 
>> >> >>> >> >> > <d...@vanderster.com>:
>> >> >>> >> >> >>
>> >> >>> >> >> >> OK
>> >> >>> >> >> >> Btw, you might need to fail to a new mgr... I'm not sure if 
>> >> >>> >> >> >> the current active will read that new config.
>> >> >>> >> >> >>
>> >> >>> >> >> >> .. dan
>> >> >>> >> >> >>
>> >> >>> >> >> >>
>> >> >>> >> >> >> On Sat, Mar 13, 2021, 4:36 PM Boris Behrens <b...@kervyn.de> 
>> >> >>> >> >> >> wrote:
>> >> >>> >> >> >>>
>> >> >>> >> >> >>> Hi,
>> >> >>> >> >> >>>
>> >> >>> >> >> >>> ok thanks. I just changed the value and rewighted 
>> >> >>> >> >> >>> everything back to 1. Now I let it sync the weekend and 
>> >> >>> >> >> >>> check how it will be on monday.
>> >> >>> >> >> >>> We tried to have the systems total storage balanced as 
>> >> >>> >> >> >>> possible. New systems will be with 8TB disks but for the 
>> >> >>> >> >> >>> exiting ones we added 16TB to offset the 4TB disks and we 
>> >> >>> >> >> >>> needed a lot of storage fast, because of a DC move. If you 
>> >> >>> >> >> >>> have any recommendations I would be happy to hear them.
>> >> >>> >> >> >>>
>> >> >>> >> >> >>> Cheers
>> >> >>> >> >> >>>  Boris
>> >> >>> >> >> >>>
>> >> >>> >> >> >>> Am Sa., 13. März 2021 um 16:20 Uhr schrieb Dan van der Ster 
>> >> >>> >> >> >>> <d...@vanderster.com>:
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>> Thanks.
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>> Decreasing the max deviation to 2 or 1 should help in your 
>> >> >>> >> >> >>>> case. This option controls when the balancer stops trying 
>> >> >>> >> >> >>>> to move PGs around -- by default it stops when the 
>> >> >>> >> >> >>>> deviation from the mean is 5. Yes this is too large IMO -- 
>> >> >>> >> >> >>>> all of our clusters have this set to 1.
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>> And given that you have some OSDs with more than 200 PGs, 
>> >> >>> >> >> >>>> you definitely shouldn't increase the num PGs.
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>> But anyway with your mixed device sizes it might be 
>> >> >>> >> >> >>>> challenging to make a perfectly uniform distribution. Give 
>> >> >>> >> >> >>>> it a try with 1 though, and let us know how it goes.
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>> .. Dan
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>>
>> >> >>> >> >> >>>> On Sat, Mar 13, 2021, 4:11 PM Boris Behrens 
>> >> >>> >> >> >>>> <b...@kervyn.de> wrote:
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>> Hi Dan,
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>> upmap_max_deviation is default (5) in our cluster. Is 1 
>> >> >>> >> >> >>>>> the recommended deviation?
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>> I added the whole ceph osd df tree, (I need to remove 
>> >> >>> >> >> >>>>> some OSDs and readd them as bluestore with SSD, so 69, 73 
>> >> >>> >> >> >>>>> and 82 are a bit off now. I also reweighted to try to get 
>> >> >>> >> >> >>>>> the %USE mitigated).
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>> I will increase the mgr debugging to see what is the 
>> >> >>> >> >> >>>>> problem.
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>> [root@s3db1 ~]# ceph osd df tree
>> >> >>> >> >> >>>>> ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE DATA    OMAP 
>> >> >>> >> >> >>>>>    META    AVAIL   %USE  VAR  PGS STATUS TYPE NAME
>> >> >>> >> >> >>>>>  -1       673.54224        - 659 TiB 491 TiB 464 TiB  96 
>> >> >>> >> >> >>>>> GiB 1.2 TiB 168 TiB 74.57 1.00   -        root default
>> >> >>> >> >> >>>>>  -2        58.30331        -  44 TiB  22 TiB  17 TiB 5.7 
>> >> >>> >> >> >>>>> GiB  38 GiB  22 TiB 49.82 0.67   -            host s3db1
>> >> >>> >> >> >>>>>  23   hdd  14.65039  1.00000  15 TiB 1.8 TiB 1.7 TiB 156 
>> >> >>> >> >> >>>>> MiB 4.4 GiB  13 TiB 12.50 0.17 101     up         osd.23
>> >> >>> >> >> >>>>>  69   hdd  14.55269        0     0 B     0 B     0 B     
>> >> >>> >> >> >>>>> 0 B     0 B     0 B     0    0  11     up         osd.69
>> >> >>> >> >> >>>>>  73   hdd  14.55269  1.00000  15 TiB  10 TiB  10 TiB 6.1 
>> >> >>> >> >> >>>>> MiB  33 GiB 4.2 TiB 71.15 0.95 107     up         osd.73
>> >> >>> >> >> >>>>>  79   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 747 GiB 2.0 
>> >> >>> >> >> >>>>> GiB     0 B 747 GiB 79.94 1.07  52     up         osd.79
>> >> >>> >> >> >>>>>  80   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1.0 TiB 1.9 
>> >> >>> >> >> >>>>> GiB     0 B 1.0 TiB 71.61 0.96  58     up         osd.80
>> >> >>> >> >> >>>>>  81   hdd   3.63689  1.00000 3.6 TiB 2.2 TiB 1.5 TiB 1.1 
>> >> >>> >> >> >>>>> GiB     0 B 1.5 TiB 60.07 0.81  55     up         osd.81
>> >> >>> >> >> >>>>>  82   hdd   3.63689  1.00000 3.6 TiB 1.9 TiB 1.7 TiB 536 
>> >> >>> >> >> >>>>> MiB     0 B 1.7 TiB 52.68 0.71  30     up         osd.82
>> >> >>> >> >> >>>>> -11        50.94173        -  51 TiB  38 TiB  38 TiB 3.7 
>> >> >>> >> >> >>>>> GiB 100 GiB  13 TiB 74.69 1.00   -            host s3db10
>> >> >>> >> >> >>>>>  63   hdd   7.27739  1.00000 7.3 TiB 5.5 TiB 5.5 TiB 616 
>> >> >>> >> >> >>>>> MiB  14 GiB 1.7 TiB 76.04 1.02  92     up         osd.63
>> >> >>> >> >> >>>>>  64   hdd   7.27739  1.00000 7.3 TiB 5.5 TiB 5.5 TiB 820 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.8 TiB 75.54 1.01 101     up         osd.64
>> >> >>> >> >> >>>>>  65   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB 5.3 TiB 109 
>> >> >>> >> >> >>>>> MiB  14 GiB 2.0 TiB 73.17 0.98 105     up         osd.65
>> >> >>> >> >> >>>>>  66   hdd   7.27739  1.00000 7.3 TiB 5.8 TiB 5.8 TiB 423 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.4 TiB 80.38 1.08  98     up         osd.66
>> >> >>> >> >> >>>>>  67   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB 5.1 TiB 572 
>> >> >>> >> >> >>>>> MiB  14 GiB 2.2 TiB 70.10 0.94 100     up         osd.67
>> >> >>> >> >> >>>>>  68   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB 5.3 TiB 630 
>> >> >>> >> >> >>>>> MiB  13 GiB 2.0 TiB 72.88 0.98 107     up         osd.68
>> >> >>> >> >> >>>>>  70   hdd   7.27739  1.00000 7.3 TiB 5.4 TiB 5.4 TiB 648 
>> >> >>> >> >> >>>>> MiB  14 GiB 1.8 TiB 74.73 1.00 102     up         osd.70
>> >> >>> >> >> >>>>> -12        50.99052        -  51 TiB  39 TiB  39 TiB 2.9 
>> >> >>> >> >> >>>>> GiB  99 GiB  12 TiB 77.24 1.04   -            host s3db11
>> >> >>> >> >> >>>>>  46   hdd   7.27739  1.00000 7.3 TiB 5.7 TiB 5.7 TiB 102 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.5 TiB 78.91 1.06  97     up         osd.46
>> >> >>> >> >> >>>>>  47   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB 5.2 TiB  61 
>> >> >>> >> >> >>>>> MiB  13 GiB 2.1 TiB 71.47 0.96  96     up         osd.47
>> >> >>> >> >> >>>>>  48   hdd   7.27739  1.00000 7.3 TiB 6.1 TiB 6.1 TiB 853 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.2 TiB 83.46 1.12 109     up         osd.48
>> >> >>> >> >> >>>>>  49   hdd   7.27739  1.00000 7.3 TiB 5.7 TiB 5.7 TiB 708 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.5 TiB 78.96 1.06  98     up         osd.49
>> >> >>> >> >> >>>>>  50   hdd   7.27739  1.00000 7.3 TiB 5.9 TiB 5.8 TiB 472 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.4 TiB 80.40 1.08 102     up         osd.50
>> >> >>> >> >> >>>>>  51   hdd   7.27739  1.00000 7.3 TiB 5.9 TiB 5.9 TiB 729 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.3 TiB 81.70 1.10 110     up         osd.51
>> >> >>> >> >> >>>>>  72   hdd   7.32619  1.00000 7.3 TiB 4.8 TiB 4.8 TiB  91 
>> >> >>> >> >> >>>>> MiB  12 GiB 2.5 TiB 65.82 0.88  89     up         osd.72
>> >> >>> >> >> >>>>> -37        58.55478        -  59 TiB  46 TiB  46 TiB 5.0 
>> >> >>> >> >> >>>>> GiB 124 GiB  12 TiB 79.04 1.06   -            host s3db12
>> >> >>> >> >> >>>>>  19   hdd   3.68750  1.00000 3.7 TiB 3.1 TiB 3.1 TiB 462 
>> >> >>> >> >> >>>>> MiB 8.2 GiB 559 GiB 85.18 1.14  55     up         osd.19
>> >> >>> >> >> >>>>>  71   hdd   3.68750  1.00000 3.7 TiB 2.9 TiB 2.8 TiB 3.9 
>> >> >>> >> >> >>>>> MiB 7.8 GiB 825 GiB 78.14 1.05  50     up         osd.71
>> >> >>> >> >> >>>>>  75   hdd   3.68750  1.00000 3.7 TiB 3.1 TiB 3.1 TiB 576 
>> >> >>> >> >> >>>>> MiB 8.3 GiB 555 GiB 85.29 1.14  57     up         osd.75
>> >> >>> >> >> >>>>>  76   hdd   3.68750  1.00000 3.7 TiB 3.2 TiB 3.1 TiB 239 
>> >> >>> >> >> >>>>> MiB 9.3 GiB 501 GiB 86.73 1.16  50     up         osd.76
>> >> >>> >> >> >>>>>  77   hdd  14.60159  1.00000  15 TiB  11 TiB  11 TiB 880 
>> >> >>> >> >> >>>>> MiB  30 GiB 3.6 TiB 75.57 1.01 202     up         osd.77
>> >> >>> >> >> >>>>>  78   hdd  14.60159  1.00000  15 TiB  11 TiB  11 TiB 1.0 
>> >> >>> >> >> >>>>> GiB  30 GiB 3.4 TiB 76.65 1.03 196     up         osd.78
>> >> >>> >> >> >>>>>  83   hdd  14.60159  1.00000  15 TiB  12 TiB  12 TiB 1.8 
>> >> >>> >> >> >>>>> GiB  31 GiB 2.9 TiB 80.04 1.07 223     up         osd.83
>> >> >>> >> >> >>>>>  -3        58.49872        -  58 TiB  43 TiB  38 TiB 8.1 
>> >> >>> >> >> >>>>> GiB  91 GiB  16 TiB 73.15 0.98   -            host s3db2
>> >> >>> >> >> >>>>>   1   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 3.1 
>> >> >>> >> >> >>>>> GiB  38 GiB 3.6 TiB 75.52 1.01 194     up         osd.1
>> >> >>> >> >> >>>>>   3   hdd   3.63689  1.00000 3.6 TiB 2.2 TiB 1.4 TiB 418 
>> >> >>> >> >> >>>>> MiB     0 B 1.4 TiB 60.94 0.82  52     up         osd.3
>> >> >>> >> >> >>>>>   4   hdd   3.63689  0.89999 3.6 TiB 3.2 TiB 401 GiB 845 
>> >> >>> >> >> >>>>> MiB     0 B 401 GiB 89.23 1.20  53     up         osd.4
>> >> >>> >> >> >>>>>   5   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB 1.3 TiB 437 
>> >> >>> >> >> >>>>> MiB     0 B 1.3 TiB 62.88 0.84  51     up         osd.5
>> >> >>> >> >> >>>>>   6   hdd   3.63689  1.00000 3.6 TiB 2.0 TiB 1.7 TiB 1.8 
>> >> >>> >> >> >>>>> GiB     0 B 1.7 TiB 54.51 0.73  47     up         osd.6
>> >> >>> >> >> >>>>>   7   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 493 
>> >> >>> >> >> >>>>> MiB  26 GiB 3.8 TiB 73.90 0.99 185     up         osd.7
>> >> >>> >> >> >>>>>  74   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 1.1 
>> >> >>> >> >> >>>>> GiB  27 GiB 3.5 TiB 76.27 1.02 208     up         osd.74
>> >> >>> >> >> >>>>>  -4        58.49872        -  58 TiB  43 TiB  37 TiB  33 
>> >> >>> >> >> >>>>> GiB  86 GiB  15 TiB 74.05 0.99   -            host s3db3
>> >> >>> >> >> >>>>>   2   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 850 
>> >> >>> >> >> >>>>> MiB  26 GiB 4.0 TiB 72.78 0.98 203     up         osd.2
>> >> >>> >> >> >>>>>   9   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 8.3 
>> >> >>> >> >> >>>>> GiB  33 GiB 3.6 TiB 75.62 1.01 189     up         osd.9
>> >> >>> >> >> >>>>>  10   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 663 
>> >> >>> >> >> >>>>> MiB  28 GiB 3.5 TiB 76.34 1.02 211     up         osd.10
>> >> >>> >> >> >>>>>  12   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB 1.2 TiB 633 
>> >> >>> >> >> >>>>> MiB     0 B 1.2 TiB 66.22 0.89  44     up         osd.12
>> >> >>> >> >> >>>>>  13   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 720 GiB 2.3 
>> >> >>> >> >> >>>>> GiB     0 B 720 GiB 80.66 1.08  66     up         osd.13
>> >> >>> >> >> >>>>>  14   hdd   3.63689  1.00000 3.6 TiB 3.1 TiB 552 GiB  18 
>> >> >>> >> >> >>>>> GiB     0 B 552 GiB 85.18 1.14  60     up         osd.14
>> >> >>> >> >> >>>>>  15   hdd   3.63689  1.00000 3.6 TiB 2.0 TiB 1.7 TiB 2.1 
>> >> >>> >> >> >>>>> GiB     0 B 1.7 TiB 53.72 0.72  44     up         osd.15
>> >> >>> >> >> >>>>>  -5        58.49872        -  58 TiB  45 TiB  37 TiB 7.2 
>> >> >>> >> >> >>>>> GiB  99 GiB  14 TiB 76.37 1.02   -            host s3db4
>> >> >>> >> >> >>>>>  11   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 897 
>> >> >>> >> >> >>>>> MiB  28 GiB 2.8 TiB 81.15 1.09 205     up         osd.11
>> >> >>> >> >> >>>>>  17   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 1.2 
>> >> >>> >> >> >>>>> GiB  27 GiB 3.6 TiB 75.38 1.01 211     up         osd.17
>> >> >>> >> >> >>>>>  18   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 965 
>> >> >>> >> >> >>>>> MiB  44 GiB 4.0 TiB 72.86 0.98 188     up         osd.18
>> >> >>> >> >> >>>>>  20   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 796 GiB 529 
>> >> >>> >> >> >>>>> MiB     0 B 796 GiB 78.63 1.05  66     up         osd.20
>> >> >>> >> >> >>>>>  21   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1.1 TiB 2.1 
>> >> >>> >> >> >>>>> GiB     0 B 1.1 TiB 70.32 0.94  47     up         osd.21
>> >> >>> >> >> >>>>>  22   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 802 GiB 882 
>> >> >>> >> >> >>>>> MiB     0 B 802 GiB 78.47 1.05  58     up         osd.22
>> >> >>> >> >> >>>>>  24   hdd   3.63689  1.00000 3.6 TiB 2.8 TiB 856 GiB 645 
>> >> >>> >> >> >>>>> MiB     0 B 856 GiB 77.01 1.03  47     up         osd.24
>> >> >>> >> >> >>>>>  -6        58.89636        -  59 TiB  44 TiB  44 TiB 2.4 
>> >> >>> >> >> >>>>> GiB 111 GiB  15 TiB 75.22 1.01   -            host s3db5
>> >> >>> >> >> >>>>>   0   hdd   3.73630  1.00000 3.7 TiB 2.4 TiB 2.3 TiB  70 
>> >> >>> >> >> >>>>> MiB 6.6 GiB 1.3 TiB 65.00 0.87  48     up         osd.0
>> >> >>> >> >> >>>>>  25   hdd   3.73630  1.00000 3.7 TiB 2.4 TiB 2.3 TiB 5.3 
>> >> >>> >> >> >>>>> MiB 6.6 GiB 1.4 TiB 63.86 0.86  41     up         osd.25
>> >> >>> >> >> >>>>>  26   hdd   3.73630  1.00000 3.7 TiB 2.9 TiB 2.8 TiB 181 
>> >> >>> >> >> >>>>> MiB 7.6 GiB 862 GiB 77.47 1.04  48     up         osd.26
>> >> >>> >> >> >>>>>  27   hdd   3.73630  1.00000 3.7 TiB 2.3 TiB 2.2 TiB 7.0 
>> >> >>> >> >> >>>>> MiB 6.1 GiB 1.5 TiB 61.00 0.82  48     up         osd.27
>> >> >>> >> >> >>>>>  28   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 937 
>> >> >>> >> >> >>>>> MiB  30 GiB 2.8 TiB 81.19 1.09 203     up         osd.28
>> >> >>> >> >> >>>>>  29   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 536 
>> >> >>> >> >> >>>>> MiB  26 GiB 3.8 TiB 73.95 0.99 200     up         osd.29
>> >> >>> >> >> >>>>>  30   hdd  14.65039  1.00000  15 TiB  12 TiB  11 TiB 744 
>> >> >>> >> >> >>>>> MiB  28 GiB 3.1 TiB 79.07 1.06 207     up         osd.30
>> >> >>> >> >> >>>>>  -7        58.89636        -  59 TiB  44 TiB  44 TiB  14 
>> >> >>> >> >> >>>>> GiB 122 GiB  14 TiB 75.41 1.01   -            host s3db6
>> >> >>> >> >> >>>>>  32   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB 3.0 TiB  16 
>> >> >>> >> >> >>>>> MiB 8.2 GiB 622 GiB 83.74 1.12  65     up         osd.32
>> >> >>> >> >> >>>>>  33   hdd   3.73630  0.79999 3.7 TiB 3.0 TiB 2.9 TiB  14 
>> >> >>> >> >> >>>>> MiB 8.1 GiB 740 GiB 80.67 1.08  52     up         osd.33
>> >> >>> >> >> >>>>>  34   hdd   3.73630  0.79999 3.7 TiB 2.9 TiB 2.8 TiB 449 
>> >> >>> >> >> >>>>> MiB 7.7 GiB 877 GiB 77.08 1.03  52     up         osd.34
>> >> >>> >> >> >>>>>  35   hdd   3.73630  0.79999 3.7 TiB 2.3 TiB 2.2 TiB 133 
>> >> >>> >> >> >>>>> MiB 7.0 GiB 1.4 TiB 62.18 0.83  42     up         osd.35
>> >> >>> >> >> >>>>>  36   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 544 
>> >> >>> >> >> >>>>> MiB  26 GiB 4.0 TiB 72.98 0.98 220     up         osd.36
>> >> >>> >> >> >>>>>  37   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB  11 
>> >> >>> >> >> >>>>> GiB  38 GiB 3.6 TiB 75.30 1.01 200     up         osd.37
>> >> >>> >> >> >>>>>  38   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 1.2 
>> >> >>> >> >> >>>>> GiB  28 GiB 3.3 TiB 77.43 1.04 217     up         osd.38
>> >> >>> >> >> >>>>>  -8        58.89636        -  59 TiB  47 TiB  46 TiB 3.9 
>> >> >>> >> >> >>>>> GiB 116 GiB  12 TiB 78.98 1.06   -            host s3db7
>> >> >>> >> >> >>>>>  39   hdd   3.73630  1.00000 3.7 TiB 3.2 TiB 3.2 TiB  19 
>> >> >>> >> >> >>>>> MiB 8.5 GiB 499 GiB 86.96 1.17  43     up         osd.39
>> >> >>> >> >> >>>>>  40   hdd   3.73630  1.00000 3.7 TiB 2.6 TiB 2.5 TiB 144 
>> >> >>> >> >> >>>>> MiB 7.0 GiB 1.2 TiB 68.33 0.92  39     up         osd.40
>> >> >>> >> >> >>>>>  41   hdd   3.73630  1.00000 3.7 TiB 3.0 TiB 2.9 TiB 218 
>> >> >>> >> >> >>>>> MiB 7.9 GiB 732 GiB 80.86 1.08  64     up         osd.41
>> >> >>> >> >> >>>>>  42   hdd   3.73630  1.00000 3.7 TiB 2.5 TiB 2.4 TiB 594 
>> >> >>> >> >> >>>>> MiB 7.0 GiB 1.2 TiB 67.97 0.91  50     up         osd.42
>> >> >>> >> >> >>>>>  43   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 564 
>> >> >>> >> >> >>>>> MiB  28 GiB 2.9 TiB 80.32 1.08 213     up         osd.43
>> >> >>> >> >> >>>>>  44   hdd  14.65039  1.00000  15 TiB  12 TiB  11 TiB 1.3 
>> >> >>> >> >> >>>>> GiB  28 GiB 3.1 TiB 78.59 1.05 198     up         osd.44
>> >> >>> >> >> >>>>>  45   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 1.2 
>> >> >>> >> >> >>>>> GiB  30 GiB 2.8 TiB 81.05 1.09 214     up         osd.45
>> >> >>> >> >> >>>>>  -9        51.28331        -  51 TiB  41 TiB  41 TiB 4.9 
>> >> >>> >> >> >>>>> GiB 108 GiB  10 TiB 79.75 1.07   -            host s3db8
>> >> >>> >> >> >>>>>   8   hdd   7.32619  1.00000 7.3 TiB 5.8 TiB 5.8 TiB 472 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.5 TiB 79.68 1.07  99     up         osd.8
>> >> >>> >> >> >>>>>  16   hdd   7.32619  1.00000 7.3 TiB 5.9 TiB 5.8 TiB 785 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.4 TiB 80.25 1.08  97     up         osd.16
>> >> >>> >> >> >>>>>  31   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB 5.5 TiB 438 
>> >> >>> >> >> >>>>> MiB  14 GiB 1.8 TiB 75.36 1.01  87     up         osd.31
>> >> >>> >> >> >>>>>  52   hdd   7.32619  1.00000 7.3 TiB 5.7 TiB 5.7 TiB 844 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.6 TiB 78.19 1.05 113     up         osd.52
>> >> >>> >> >> >>>>>  53   hdd   7.32619  1.00000 7.3 TiB 6.2 TiB 6.1 TiB 792 
>> >> >>> >> >> >>>>> MiB  18 GiB 1.1 TiB 84.46 1.13 109     up         osd.53
>> >> >>> >> >> >>>>>  54   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB 5.6 TiB 959 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.7 TiB 76.73 1.03 115     up         osd.54
>> >> >>> >> >> >>>>>  55   hdd   7.32619  1.00000 7.3 TiB 6.1 TiB 6.1 TiB 699 
>> >> >>> >> >> >>>>> MiB  16 GiB 1.2 TiB 83.56 1.12 122     up         osd.55
>> >> >>> >> >> >>>>> -10        51.28331        -  51 TiB  39 TiB  39 TiB 4.7 
>> >> >>> >> >> >>>>> GiB 100 GiB  12 TiB 76.05 1.02   -            host s3db9
>> >> >>> >> >> >>>>>  56   hdd   7.32619  1.00000 7.3 TiB 5.2 TiB 5.2 TiB 840 
>> >> >>> >> >> >>>>> MiB  13 GiB 2.1 TiB 71.06 0.95 105     up         osd.56
>> >> >>> >> >> >>>>>  57   hdd   7.32619  1.00000 7.3 TiB 6.1 TiB 6.0 TiB 1.0 
>> >> >>> >> >> >>>>> GiB  16 GiB 1.2 TiB 83.17 1.12 102     up         osd.57
>> >> >>> >> >> >>>>>  58   hdd   7.32619  1.00000 7.3 TiB 6.0 TiB 5.9 TiB  43 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.4 TiB 81.56 1.09 105     up         osd.58
>> >> >>> >> >> >>>>>  59   hdd   7.32619  1.00000 7.3 TiB 5.9 TiB 5.9 TiB 429 
>> >> >>> >> >> >>>>> MiB  15 GiB 1.4 TiB 80.64 1.08  94     up         osd.59
>> >> >>> >> >> >>>>>  60   hdd   7.32619  1.00000 7.3 TiB 5.4 TiB 5.3 TiB 226 
>> >> >>> >> >> >>>>> MiB  14 GiB 2.0 TiB 73.25 0.98 101     up         osd.60
>> >> >>> >> >> >>>>>  61   hdd   7.32619  1.00000 7.3 TiB 4.8 TiB 4.8 TiB 1.1 
>> >> >>> >> >> >>>>> GiB  12 GiB 2.5 TiB 65.84 0.88 103     up         osd.61
>> >> >>> >> >> >>>>>  62   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB 5.6 TiB 1.0 
>> >> >>> >> >> >>>>> GiB  15 GiB 1.7 TiB 76.83 1.03 126     up         osd.62
>> >> >>> >> >> >>>>>                        TOTAL 674 TiB 501 TiB 473 TiB  96 
>> >> >>> >> >> >>>>> GiB 1.2 TiB 173 TiB 74.57
>> >> >>> >> >> >>>>> MIN/MAX VAR: 0.17/1.20  STDDEV: 10.25
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>> Am Sa., 13. März 2021 um 15:57 Uhr schrieb Dan van der 
>> >> >>> >> >> >>>>> Ster <d...@vanderster.com>:
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>> No, increasing num PGs won't help substantially.
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>> Can you share the entire output of ceph osd df tree ?
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>> Did you already set
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>>   ceph config set mgr mgr/balancer/upmap_max_deviation 1
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>> ??
>> >> >>> >> >> >>>>>> And I recommend debug_mgr 4/5 so you can see some basic 
>> >> >>> >> >> >>>>>> upmap balancer logging.
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>> .. Dan
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>>
>> >> >>> >> >> >>>>>> On Sat, Mar 13, 2021, 3:49 PM Boris Behrens 
>> >> >>> >> >> >>>>>> <b...@kervyn.de> wrote:
>> >> >>> >> >> >>>>>>>
>> >> >>> >> >> >>>>>>> Hello people,
>> >> >>> >> >> >>>>>>>
>> >> >>> >> >> >>>>>>> I am still struggeling with the balancer
>> >> >>> >> >> >>>>>>> (https://www.mail-archive.com/ceph-users@ceph.io/msg09124.html)
>> >> >>> >> >> >>>>>>> Now I've read some more and might think that I do not 
>> >> >>> >> >> >>>>>>> have enough PGs.
>> >> >>> >> >> >>>>>>> Currently I have 84OSDs and 1024PGs for the main pool 
>> >> >>> >> >> >>>>>>> (3008 total). I
>> >> >>> >> >> >>>>>>> have the autoscaler enabled, but I doesn't tell me to 
>> >> >>> >> >> >>>>>>> increase the
>> >> >>> >> >> >>>>>>> PGs.
>> >> >>> >> >> >>>>>>>
>> >> >>> >> >> >>>>>>> What do you think?
>> >> >>> >> >> >>>>>>>
>> >> >>> >> >> >>>>>>> --
>> >> >>> >> >> >>>>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich 
>> >> >>> >> >> >>>>>>> diesmal abweichend
>> >> >>> >> >> >>>>>>> im groüen Saal.
>> >> >>> >> >> >>>>>>> _______________________________________________
>> >> >>> >> >> >>>>>>> ceph-users mailing list -- ceph-users@ceph.io
>> >> >>> >> >> >>>>>>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>>
>> >> >>> >> >> >>>>> --
>> >> >>> >> >> >>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich 
>> >> >>> >> >> >>>>> diesmal abweichend im groüen Saal.
>> >> >>> >> >> >>>
>> >> >>> >> >> >>>
>> >> >>> >> >> >>>
>> >> >>> >> >> >>> --
>> >> >>> >> >> >>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal 
>> >> >>> >> >> >>> abweichend im groüen Saal.
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> > --
>> >> >>> >> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal 
>> >> >>> >> >> > abweichend im groüen Saal.
>> >> >>> >> >
>> >> >>> >> >
>> >> >>> >> >
>> >> >>> >> > --
>> >> >>> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal 
>> >> >>> >> > abweichend im groüen Saal.
>> >> >>> >
>> >> >>> >
>> >> >>> >
>> >> >>> > --
>> >> >>> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal 
>> >> >>> > abweichend im groüen Saal.
>> >> >>
>> >> >>
>> >> >>
>> >> >> --
>> >> >> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend 
>> >> >> im groüen Saal.
>> >> >
>> >> >
>> >> >
>> >> > --
>> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend 
>> >> > im groüen Saal.
>> >
>> >
>> >
>> > --
>> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im 
>> > groüen Saal.
>
>
>
> --
> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im 
> groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to