This is the pool pool 41 'ec82_pool' erasure size 10 min_size 8 crush_rule 1 object_hash rjenkins pg_num 512 pgp_num 512 last_change 63794 lfor 21731/21731 flags hashpspool,ec_overwrites stripe_width 32768 application cephfs removed_snaps [1~5]
Here is the relevant crush rule: rule ec_pool { id 1 type erasure min_size 3 max_size 10 step set_chooseleaf_tries 5 step set_choose_tries 100 step take default class hdd step choose indep 5 type host step choose indep 2 type osd step emit } Both OSD 23 and 123 are in the same host. So this change should be perfectly acceptable by the rule set. Something must be blocking the change, but i can't find anything about it in any logs. - Kári On Thu, Feb 28, 2019 at 8:07 AM Dan van der Ster <d...@vanderster.com> wrote: > Hi, > > pg-upmap-items became more strict in v12.2.11 when validating upmaps. > E.g., it now won't let you put two PGs in the same rack if the crush > rule doesn't allow it. > > Where are OSDs 23 and 123 in your cluster? What is the relevant crush rule? > > -- dan > > > On Wed, Feb 27, 2019 at 9:17 PM Kári Bertilsson <karibert...@gmail.com> > wrote: > > > > Hello > > > > I am trying to diagnose why upmap stopped working where it was > previously working fine. > > > > Trying to move pg 41.1 to 123 has no effect and seems to be ignored. > > > > # ceph osd pg-upmap-items 41.1 23 123 > > set 41.1 pg_upmap_items mapping to [23->123] > > > > No rebalacing happens and if i run it again it shows the same output > every time. > > > > I have in config > > debug mgr = 4/5 > > debug mon = 4/5 > > > > Paste from mon & mgr logs. Also output from "ceph osd dump" > > https://pastebin.com/9VrT4YcU > > > > > > I have run "ceph osd set-require-min-compat-client luminous" long time > ago. And all servers running ceph have been rebooted numerous times since > then. > > But somehow i am still seeing "min_compat_client jewel". I believe that > upmap was previously working anyway with that "jewel" line present. > > > > I see no indication in any logs why the upmap commands are being ignored. > > > > Any suggestions on how to debug further or what could be the issue ? > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com