OK thanks. Indeed "prepared 0/10 changes" means it thinks things are balanced.
Could you again share the full ceph osd df tree?

On Mon, Mar 15, 2021 at 2:54 PM Boris Behrens <b...@kervyn.de> wrote:
>
> Hi Dan,
>
> I've set the autoscaler to warn, but it actually does not warn for now. So 
> not touching it for now.
>
> this is what the log says in minute intervals:
> 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/active
> 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/sleep_interval
> 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/begin_time
> 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/end_time
> 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/begin_weekday
> 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/end_weekday
> 2021-03-15 13:51:00.971 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/pool_ids
> 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] Optimize plan 
> auto_2021-03-15_13:51:00
> 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/mode
> 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] Mode upmap, max 
> misplaced 0.050000
> 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] do_upmap
> 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/upmap_max_iterations
> 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr get_config get_config key: 
> mgr/balancer/upmap_max_deviation
> 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr[balancer] pools 
> ['eu-msg-1.rgw.data.root', 'eu-msg-1.rgw.buckets.non-ec', 
> 'eu-central-1.rgw.users.keys', 'eu-central-1.rgw.gc', 
> 'eu-central-1.rgw.buckets.data', 'eu-central-1.rgw.users.email', 
> 'eu-msg-1.rgw.gc', 'eu-central-1.rgw.usage', 'eu-msg-1.rgw.users.keys', 
> 'eu-central-1.rgw.buckets.index', 'rbd', 'eu-msg-1.rgw.log', 
> 'whitespace-again-2021-03-10_2', 'eu-msg-1.rgw.buckets.index', 
> 'eu-msg-1.rgw.meta', 'eu-central-1.rgw.log', 'default.rgw.gc', 
> 'eu-central-1.rgw.buckets.non-ec', 'eu-msg-1.rgw.usage', 
> 'whitespace-again-2021-03-10', 'fra-1.rgw.meta', 
> 'eu-central-1.rgw.users.uid', 'eu-msg-1.rgw.users.email', 
> 'fra-1.rgw.control', 'eu-msg-1.rgw.users.uid', 'eu-msg-1.rgw.control', 
> '.rgw.root', 'eu-msg-1.rgw.buckets.data', 'default.rgw.control', 
> 'fra-1.rgw.log', 'default.rgw.data.root', 'whitespace-again-2021-03-10_3', 
> 'default.rgw.log', 'eu-central-1.rgw.meta', 'eu-central-1.rgw.data.root', 
> 'default.rgw.users.uid', 'eu-central-1.rgw.control']
> 2021-03-15 13:51:01.224 7f307d5fd700  4 mgr[balancer] prepared 0/10 changes
>
> Am Mo., 15. März 2021 um 14:15 Uhr schrieb Dan van der Ster 
> <d...@vanderster.com>:
>>
>> I suggest to just disable the autoscaler until your balancing is understood.
>>
>> What does your active mgr log say (with debug_mgr 4/5), grep balancer
>> /var/log/ceph/ceph-mgr.*.log
>>
>> -- Dan
>>
>> On Mon, Mar 15, 2021 at 1:47 PM Boris Behrens <b...@kervyn.de> wrote:
>> >
>> > Hi,
>> > this unfortunally did not solve my problem. I still have some OSDs that 
>> > fill up to 85%
>> >
>> > According to the logging, the autoscaler might want to add more PGs to one 
>> > Bucken and reduce almost all other buckets to 32.
>> > 2021-03-15 12:19:58.825 7f307f601700  4 mgr[pg_autoscaler] Pool 
>> > 'eu-central-1.rgw.buckets.data' root_id -1 using 0.705080476146 of space, 
>> > bias 1.0, pg target 1974.22533321 quantized to 2048 (current 1024)
>> >
>> > Why the balancing does not happen is still nebulous to me.
>> >
>> >
>> >
>> > Am Sa., 13. März 2021 um 16:37 Uhr schrieb Dan van der Ster 
>> > <d...@vanderster.com>:
>> >>
>> >> OK
>> >> Btw, you might need to fail to a new mgr... I'm not sure if the current 
>> >> active will read that new config.
>> >>
>> >> .. dan
>> >>
>> >>
>> >> On Sat, Mar 13, 2021, 4:36 PM Boris Behrens <b...@kervyn.de> wrote:
>> >>>
>> >>> Hi,
>> >>>
>> >>> ok thanks. I just changed the value and rewighted everything back to 1. 
>> >>> Now I let it sync the weekend and check how it will be on monday.
>> >>> We tried to have the systems total storage balanced as possible. New 
>> >>> systems will be with 8TB disks but for the exiting ones we added 16TB to 
>> >>> offset the 4TB disks and we needed a lot of storage fast, because of a 
>> >>> DC move. If you have any recommendations I would be happy to hear them.
>> >>>
>> >>> Cheers
>> >>>  Boris
>> >>>
>> >>> Am Sa., 13. März 2021 um 16:20 Uhr schrieb Dan van der Ster 
>> >>> <d...@vanderster.com>:
>> >>>>
>> >>>> Thanks.
>> >>>>
>> >>>> Decreasing the max deviation to 2 or 1 should help in your case. This 
>> >>>> option controls when the balancer stops trying to move PGs around -- by 
>> >>>> default it stops when the deviation from the mean is 5. Yes this is too 
>> >>>> large IMO -- all of our clusters have this set to 1.
>> >>>>
>> >>>> And given that you have some OSDs with more than 200 PGs, you 
>> >>>> definitely shouldn't increase the num PGs.
>> >>>>
>> >>>> But anyway with your mixed device sizes it might be challenging to make 
>> >>>> a perfectly uniform distribution. Give it a try with 1 though, and let 
>> >>>> us know how it goes.
>> >>>>
>> >>>> .. Dan
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>> On Sat, Mar 13, 2021, 4:11 PM Boris Behrens <b...@kervyn.de> wrote:
>> >>>>>
>> >>>>> Hi Dan,
>> >>>>>
>> >>>>> upmap_max_deviation is default (5) in our cluster. Is 1 the 
>> >>>>> recommended deviation?
>> >>>>>
>> >>>>> I added the whole ceph osd df tree, (I need to remove some OSDs and 
>> >>>>> readd them as bluestore with SSD, so 69, 73 and 82 are a bit off now. 
>> >>>>> I also reweighted to try to get the %USE mitigated).
>> >>>>>
>> >>>>> I will increase the mgr debugging to see what is the problem.
>> >>>>>
>> >>>>> [root@s3db1 ~]# ceph osd df tree
>> >>>>> ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE DATA    OMAP    META    
>> >>>>> AVAIL   %USE  VAR  PGS STATUS TYPE NAME
>> >>>>>  -1       673.54224        - 659 TiB 491 TiB 464 TiB  96 GiB 1.2 TiB 
>> >>>>> 168 TiB 74.57 1.00   -        root default
>> >>>>>  -2        58.30331        -  44 TiB  22 TiB  17 TiB 5.7 GiB  38 GiB  
>> >>>>> 22 TiB 49.82 0.67   -            host s3db1
>> >>>>>  23   hdd  14.65039  1.00000  15 TiB 1.8 TiB 1.7 TiB 156 MiB 4.4 GiB  
>> >>>>> 13 TiB 12.50 0.17 101     up         osd.23
>> >>>>>  69   hdd  14.55269        0     0 B     0 B     0 B     0 B     0 B   
>> >>>>>   0 B     0    0  11     up         osd.69
>> >>>>>  73   hdd  14.55269  1.00000  15 TiB  10 TiB  10 TiB 6.1 MiB  33 GiB 
>> >>>>> 4.2 TiB 71.15 0.95 107     up         osd.73
>> >>>>>  79   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 747 GiB 2.0 GiB     0 B 
>> >>>>> 747 GiB 79.94 1.07  52     up         osd.79
>> >>>>>  80   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1.0 TiB 1.9 GiB     0 B 
>> >>>>> 1.0 TiB 71.61 0.96  58     up         osd.80
>> >>>>>  81   hdd   3.63689  1.00000 3.6 TiB 2.2 TiB 1.5 TiB 1.1 GiB     0 B 
>> >>>>> 1.5 TiB 60.07 0.81  55     up         osd.81
>> >>>>>  82   hdd   3.63689  1.00000 3.6 TiB 1.9 TiB 1.7 TiB 536 MiB     0 B 
>> >>>>> 1.7 TiB 52.68 0.71  30     up         osd.82
>> >>>>> -11        50.94173        -  51 TiB  38 TiB  38 TiB 3.7 GiB 100 GiB  
>> >>>>> 13 TiB 74.69 1.00   -            host s3db10
>> >>>>>  63   hdd   7.27739  1.00000 7.3 TiB 5.5 TiB 5.5 TiB 616 MiB  14 GiB 
>> >>>>> 1.7 TiB 76.04 1.02  92     up         osd.63
>> >>>>>  64   hdd   7.27739  1.00000 7.3 TiB 5.5 TiB 5.5 TiB 820 MiB  15 GiB 
>> >>>>> 1.8 TiB 75.54 1.01 101     up         osd.64
>> >>>>>  65   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB 5.3 TiB 109 MiB  14 GiB 
>> >>>>> 2.0 TiB 73.17 0.98 105     up         osd.65
>> >>>>>  66   hdd   7.27739  1.00000 7.3 TiB 5.8 TiB 5.8 TiB 423 MiB  15 GiB 
>> >>>>> 1.4 TiB 80.38 1.08  98     up         osd.66
>> >>>>>  67   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB 5.1 TiB 572 MiB  14 GiB 
>> >>>>> 2.2 TiB 70.10 0.94 100     up         osd.67
>> >>>>>  68   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB 5.3 TiB 630 MiB  13 GiB 
>> >>>>> 2.0 TiB 72.88 0.98 107     up         osd.68
>> >>>>>  70   hdd   7.27739  1.00000 7.3 TiB 5.4 TiB 5.4 TiB 648 MiB  14 GiB 
>> >>>>> 1.8 TiB 74.73 1.00 102     up         osd.70
>> >>>>> -12        50.99052        -  51 TiB  39 TiB  39 TiB 2.9 GiB  99 GiB  
>> >>>>> 12 TiB 77.24 1.04   -            host s3db11
>> >>>>>  46   hdd   7.27739  1.00000 7.3 TiB 5.7 TiB 5.7 TiB 102 MiB  15 GiB 
>> >>>>> 1.5 TiB 78.91 1.06  97     up         osd.46
>> >>>>>  47   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB 5.2 TiB  61 MiB  13 GiB 
>> >>>>> 2.1 TiB 71.47 0.96  96     up         osd.47
>> >>>>>  48   hdd   7.27739  1.00000 7.3 TiB 6.1 TiB 6.1 TiB 853 MiB  15 GiB 
>> >>>>> 1.2 TiB 83.46 1.12 109     up         osd.48
>> >>>>>  49   hdd   7.27739  1.00000 7.3 TiB 5.7 TiB 5.7 TiB 708 MiB  15 GiB 
>> >>>>> 1.5 TiB 78.96 1.06  98     up         osd.49
>> >>>>>  50   hdd   7.27739  1.00000 7.3 TiB 5.9 TiB 5.8 TiB 472 MiB  15 GiB 
>> >>>>> 1.4 TiB 80.40 1.08 102     up         osd.50
>> >>>>>  51   hdd   7.27739  1.00000 7.3 TiB 5.9 TiB 5.9 TiB 729 MiB  15 GiB 
>> >>>>> 1.3 TiB 81.70 1.10 110     up         osd.51
>> >>>>>  72   hdd   7.32619  1.00000 7.3 TiB 4.8 TiB 4.8 TiB  91 MiB  12 GiB 
>> >>>>> 2.5 TiB 65.82 0.88  89     up         osd.72
>> >>>>> -37        58.55478        -  59 TiB  46 TiB  46 TiB 5.0 GiB 124 GiB  
>> >>>>> 12 TiB 79.04 1.06   -            host s3db12
>> >>>>>  19   hdd   3.68750  1.00000 3.7 TiB 3.1 TiB 3.1 TiB 462 MiB 8.2 GiB 
>> >>>>> 559 GiB 85.18 1.14  55     up         osd.19
>> >>>>>  71   hdd   3.68750  1.00000 3.7 TiB 2.9 TiB 2.8 TiB 3.9 MiB 7.8 GiB 
>> >>>>> 825 GiB 78.14 1.05  50     up         osd.71
>> >>>>>  75   hdd   3.68750  1.00000 3.7 TiB 3.1 TiB 3.1 TiB 576 MiB 8.3 GiB 
>> >>>>> 555 GiB 85.29 1.14  57     up         osd.75
>> >>>>>  76   hdd   3.68750  1.00000 3.7 TiB 3.2 TiB 3.1 TiB 239 MiB 9.3 GiB 
>> >>>>> 501 GiB 86.73 1.16  50     up         osd.76
>> >>>>>  77   hdd  14.60159  1.00000  15 TiB  11 TiB  11 TiB 880 MiB  30 GiB 
>> >>>>> 3.6 TiB 75.57 1.01 202     up         osd.77
>> >>>>>  78   hdd  14.60159  1.00000  15 TiB  11 TiB  11 TiB 1.0 GiB  30 GiB 
>> >>>>> 3.4 TiB 76.65 1.03 196     up         osd.78
>> >>>>>  83   hdd  14.60159  1.00000  15 TiB  12 TiB  12 TiB 1.8 GiB  31 GiB 
>> >>>>> 2.9 TiB 80.04 1.07 223     up         osd.83
>> >>>>>  -3        58.49872        -  58 TiB  43 TiB  38 TiB 8.1 GiB  91 GiB  
>> >>>>> 16 TiB 73.15 0.98   -            host s3db2
>> >>>>>   1   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 3.1 GiB  38 GiB 
>> >>>>> 3.6 TiB 75.52 1.01 194     up         osd.1
>> >>>>>   3   hdd   3.63689  1.00000 3.6 TiB 2.2 TiB 1.4 TiB 418 MiB     0 B 
>> >>>>> 1.4 TiB 60.94 0.82  52     up         osd.3
>> >>>>>   4   hdd   3.63689  0.89999 3.6 TiB 3.2 TiB 401 GiB 845 MiB     0 B 
>> >>>>> 401 GiB 89.23 1.20  53     up         osd.4
>> >>>>>   5   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB 1.3 TiB 437 MiB     0 B 
>> >>>>> 1.3 TiB 62.88 0.84  51     up         osd.5
>> >>>>>   6   hdd   3.63689  1.00000 3.6 TiB 2.0 TiB 1.7 TiB 1.8 GiB     0 B 
>> >>>>> 1.7 TiB 54.51 0.73  47     up         osd.6
>> >>>>>   7   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 493 MiB  26 GiB 
>> >>>>> 3.8 TiB 73.90 0.99 185     up         osd.7
>> >>>>>  74   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 1.1 GiB  27 GiB 
>> >>>>> 3.5 TiB 76.27 1.02 208     up         osd.74
>> >>>>>  -4        58.49872        -  58 TiB  43 TiB  37 TiB  33 GiB  86 GiB  
>> >>>>> 15 TiB 74.05 0.99   -            host s3db3
>> >>>>>   2   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 850 MiB  26 GiB 
>> >>>>> 4.0 TiB 72.78 0.98 203     up         osd.2
>> >>>>>   9   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 8.3 GiB  33 GiB 
>> >>>>> 3.6 TiB 75.62 1.01 189     up         osd.9
>> >>>>>  10   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 663 MiB  28 GiB 
>> >>>>> 3.5 TiB 76.34 1.02 211     up         osd.10
>> >>>>>  12   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB 1.2 TiB 633 MiB     0 B 
>> >>>>> 1.2 TiB 66.22 0.89  44     up         osd.12
>> >>>>>  13   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 720 GiB 2.3 GiB     0 B 
>> >>>>> 720 GiB 80.66 1.08  66     up         osd.13
>> >>>>>  14   hdd   3.63689  1.00000 3.6 TiB 3.1 TiB 552 GiB  18 GiB     0 B 
>> >>>>> 552 GiB 85.18 1.14  60     up         osd.14
>> >>>>>  15   hdd   3.63689  1.00000 3.6 TiB 2.0 TiB 1.7 TiB 2.1 GiB     0 B 
>> >>>>> 1.7 TiB 53.72 0.72  44     up         osd.15
>> >>>>>  -5        58.49872        -  58 TiB  45 TiB  37 TiB 7.2 GiB  99 GiB  
>> >>>>> 14 TiB 76.37 1.02   -            host s3db4
>> >>>>>  11   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 897 MiB  28 GiB 
>> >>>>> 2.8 TiB 81.15 1.09 205     up         osd.11
>> >>>>>  17   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 1.2 GiB  27 GiB 
>> >>>>> 3.6 TiB 75.38 1.01 211     up         osd.17
>> >>>>>  18   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 965 MiB  44 GiB 
>> >>>>> 4.0 TiB 72.86 0.98 188     up         osd.18
>> >>>>>  20   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 796 GiB 529 MiB     0 B 
>> >>>>> 796 GiB 78.63 1.05  66     up         osd.20
>> >>>>>  21   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1.1 TiB 2.1 GiB     0 B 
>> >>>>> 1.1 TiB 70.32 0.94  47     up         osd.21
>> >>>>>  22   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 802 GiB 882 MiB     0 B 
>> >>>>> 802 GiB 78.47 1.05  58     up         osd.22
>> >>>>>  24   hdd   3.63689  1.00000 3.6 TiB 2.8 TiB 856 GiB 645 MiB     0 B 
>> >>>>> 856 GiB 77.01 1.03  47     up         osd.24
>> >>>>>  -6        58.89636        -  59 TiB  44 TiB  44 TiB 2.4 GiB 111 GiB  
>> >>>>> 15 TiB 75.22 1.01   -            host s3db5
>> >>>>>   0   hdd   3.73630  1.00000 3.7 TiB 2.4 TiB 2.3 TiB  70 MiB 6.6 GiB 
>> >>>>> 1.3 TiB 65.00 0.87  48     up         osd.0
>> >>>>>  25   hdd   3.73630  1.00000 3.7 TiB 2.4 TiB 2.3 TiB 5.3 MiB 6.6 GiB 
>> >>>>> 1.4 TiB 63.86 0.86  41     up         osd.25
>> >>>>>  26   hdd   3.73630  1.00000 3.7 TiB 2.9 TiB 2.8 TiB 181 MiB 7.6 GiB 
>> >>>>> 862 GiB 77.47 1.04  48     up         osd.26
>> >>>>>  27   hdd   3.73630  1.00000 3.7 TiB 2.3 TiB 2.2 TiB 7.0 MiB 6.1 GiB 
>> >>>>> 1.5 TiB 61.00 0.82  48     up         osd.27
>> >>>>>  28   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 937 MiB  30 GiB 
>> >>>>> 2.8 TiB 81.19 1.09 203     up         osd.28
>> >>>>>  29   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 536 MiB  26 GiB 
>> >>>>> 3.8 TiB 73.95 0.99 200     up         osd.29
>> >>>>>  30   hdd  14.65039  1.00000  15 TiB  12 TiB  11 TiB 744 MiB  28 GiB 
>> >>>>> 3.1 TiB 79.07 1.06 207     up         osd.30
>> >>>>>  -7        58.89636        -  59 TiB  44 TiB  44 TiB  14 GiB 122 GiB  
>> >>>>> 14 TiB 75.41 1.01   -            host s3db6
>> >>>>>  32   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB 3.0 TiB  16 MiB 8.2 GiB 
>> >>>>> 622 GiB 83.74 1.12  65     up         osd.32
>> >>>>>  33   hdd   3.73630  0.79999 3.7 TiB 3.0 TiB 2.9 TiB  14 MiB 8.1 GiB 
>> >>>>> 740 GiB 80.67 1.08  52     up         osd.33
>> >>>>>  34   hdd   3.73630  0.79999 3.7 TiB 2.9 TiB 2.8 TiB 449 MiB 7.7 GiB 
>> >>>>> 877 GiB 77.08 1.03  52     up         osd.34
>> >>>>>  35   hdd   3.73630  0.79999 3.7 TiB 2.3 TiB 2.2 TiB 133 MiB 7.0 GiB 
>> >>>>> 1.4 TiB 62.18 0.83  42     up         osd.35
>> >>>>>  36   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 544 MiB  26 GiB 
>> >>>>> 4.0 TiB 72.98 0.98 220     up         osd.36
>> >>>>>  37   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB  11 GiB  38 GiB 
>> >>>>> 3.6 TiB 75.30 1.01 200     up         osd.37
>> >>>>>  38   hdd  14.65039  1.00000  15 TiB  11 TiB  11 TiB 1.2 GiB  28 GiB 
>> >>>>> 3.3 TiB 77.43 1.04 217     up         osd.38
>> >>>>>  -8        58.89636        -  59 TiB  47 TiB  46 TiB 3.9 GiB 116 GiB  
>> >>>>> 12 TiB 78.98 1.06   -            host s3db7
>> >>>>>  39   hdd   3.73630  1.00000 3.7 TiB 3.2 TiB 3.2 TiB  19 MiB 8.5 GiB 
>> >>>>> 499 GiB 86.96 1.17  43     up         osd.39
>> >>>>>  40   hdd   3.73630  1.00000 3.7 TiB 2.6 TiB 2.5 TiB 144 MiB 7.0 GiB 
>> >>>>> 1.2 TiB 68.33 0.92  39     up         osd.40
>> >>>>>  41   hdd   3.73630  1.00000 3.7 TiB 3.0 TiB 2.9 TiB 218 MiB 7.9 GiB 
>> >>>>> 732 GiB 80.86 1.08  64     up         osd.41
>> >>>>>  42   hdd   3.73630  1.00000 3.7 TiB 2.5 TiB 2.4 TiB 594 MiB 7.0 GiB 
>> >>>>> 1.2 TiB 67.97 0.91  50     up         osd.42
>> >>>>>  43   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 564 MiB  28 GiB 
>> >>>>> 2.9 TiB 80.32 1.08 213     up         osd.43
>> >>>>>  44   hdd  14.65039  1.00000  15 TiB  12 TiB  11 TiB 1.3 GiB  28 GiB 
>> >>>>> 3.1 TiB 78.59 1.05 198     up         osd.44
>> >>>>>  45   hdd  14.65039  1.00000  15 TiB  12 TiB  12 TiB 1.2 GiB  30 GiB 
>> >>>>> 2.8 TiB 81.05 1.09 214     up         osd.45
>> >>>>>  -9        51.28331        -  51 TiB  41 TiB  41 TiB 4.9 GiB 108 GiB  
>> >>>>> 10 TiB 79.75 1.07   -            host s3db8
>> >>>>>   8   hdd   7.32619  1.00000 7.3 TiB 5.8 TiB 5.8 TiB 472 MiB  15 GiB 
>> >>>>> 1.5 TiB 79.68 1.07  99     up         osd.8
>> >>>>>  16   hdd   7.32619  1.00000 7.3 TiB 5.9 TiB 5.8 TiB 785 MiB  15 GiB 
>> >>>>> 1.4 TiB 80.25 1.08  97     up         osd.16
>> >>>>>  31   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB 5.5 TiB 438 MiB  14 GiB 
>> >>>>> 1.8 TiB 75.36 1.01  87     up         osd.31
>> >>>>>  52   hdd   7.32619  1.00000 7.3 TiB 5.7 TiB 5.7 TiB 844 MiB  15 GiB 
>> >>>>> 1.6 TiB 78.19 1.05 113     up         osd.52
>> >>>>>  53   hdd   7.32619  1.00000 7.3 TiB 6.2 TiB 6.1 TiB 792 MiB  18 GiB 
>> >>>>> 1.1 TiB 84.46 1.13 109     up         osd.53
>> >>>>>  54   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB 5.6 TiB 959 MiB  15 GiB 
>> >>>>> 1.7 TiB 76.73 1.03 115     up         osd.54
>> >>>>>  55   hdd   7.32619  1.00000 7.3 TiB 6.1 TiB 6.1 TiB 699 MiB  16 GiB 
>> >>>>> 1.2 TiB 83.56 1.12 122     up         osd.55
>> >>>>> -10        51.28331        -  51 TiB  39 TiB  39 TiB 4.7 GiB 100 GiB  
>> >>>>> 12 TiB 76.05 1.02   -            host s3db9
>> >>>>>  56   hdd   7.32619  1.00000 7.3 TiB 5.2 TiB 5.2 TiB 840 MiB  13 GiB 
>> >>>>> 2.1 TiB 71.06 0.95 105     up         osd.56
>> >>>>>  57   hdd   7.32619  1.00000 7.3 TiB 6.1 TiB 6.0 TiB 1.0 GiB  16 GiB 
>> >>>>> 1.2 TiB 83.17 1.12 102     up         osd.57
>> >>>>>  58   hdd   7.32619  1.00000 7.3 TiB 6.0 TiB 5.9 TiB  43 MiB  15 GiB 
>> >>>>> 1.4 TiB 81.56 1.09 105     up         osd.58
>> >>>>>  59   hdd   7.32619  1.00000 7.3 TiB 5.9 TiB 5.9 TiB 429 MiB  15 GiB 
>> >>>>> 1.4 TiB 80.64 1.08  94     up         osd.59
>> >>>>>  60   hdd   7.32619  1.00000 7.3 TiB 5.4 TiB 5.3 TiB 226 MiB  14 GiB 
>> >>>>> 2.0 TiB 73.25 0.98 101     up         osd.60
>> >>>>>  61   hdd   7.32619  1.00000 7.3 TiB 4.8 TiB 4.8 TiB 1.1 GiB  12 GiB 
>> >>>>> 2.5 TiB 65.84 0.88 103     up         osd.61
>> >>>>>  62   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB 5.6 TiB 1.0 GiB  15 GiB 
>> >>>>> 1.7 TiB 76.83 1.03 126     up         osd.62
>> >>>>>                        TOTAL 674 TiB 501 TiB 473 TiB  96 GiB 1.2 TiB 
>> >>>>> 173 TiB 74.57
>> >>>>> MIN/MAX VAR: 0.17/1.20  STDDEV: 10.25
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>> Am Sa., 13. März 2021 um 15:57 Uhr schrieb Dan van der Ster 
>> >>>>> <d...@vanderster.com>:
>> >>>>>>
>> >>>>>> No, increasing num PGs won't help substantially.
>> >>>>>>
>> >>>>>> Can you share the entire output of ceph osd df tree ?
>> >>>>>>
>> >>>>>> Did you already set
>> >>>>>>
>> >>>>>>   ceph config set mgr mgr/balancer/upmap_max_deviation 1
>> >>>>>>
>> >>>>>>
>> >>>>>> ??
>> >>>>>> And I recommend debug_mgr 4/5 so you can see some basic upmap 
>> >>>>>> balancer logging.
>> >>>>>>
>> >>>>>> .. Dan
>> >>>>>>
>> >>>>>>
>> >>>>>>
>> >>>>>>
>> >>>>>>
>> >>>>>>
>> >>>>>> On Sat, Mar 13, 2021, 3:49 PM Boris Behrens <b...@kervyn.de> wrote:
>> >>>>>>>
>> >>>>>>> Hello people,
>> >>>>>>>
>> >>>>>>> I am still struggeling with the balancer
>> >>>>>>> (https://www.mail-archive.com/ceph-users@ceph.io/msg09124.html)
>> >>>>>>> Now I've read some more and might think that I do not have enough 
>> >>>>>>> PGs.
>> >>>>>>> Currently I have 84OSDs and 1024PGs for the main pool (3008 total). I
>> >>>>>>> have the autoscaler enabled, but I doesn't tell me to increase the
>> >>>>>>> PGs.
>> >>>>>>>
>> >>>>>>> What do you think?
>> >>>>>>>
>> >>>>>>> --
>> >>>>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend
>> >>>>>>> im groüen Saal.
>> >>>>>>> _______________________________________________
>> >>>>>>> ceph-users mailing list -- ceph-users@ceph.io
>> >>>>>>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>> --
>> >>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend 
>> >>>>> im groüen Saal.
>> >>>
>> >>>
>> >>>
>> >>> --
>> >>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im 
>> >>> groüen Saal.
>> >
>> >
>> >
>> > --
>> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im 
>> > groüen Saal.
>
>
>
> --
> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im 
> groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to