Hi,

looks like it fails to adjust the number of weight set entries when moving
the entries. The good news is that this is 100% reproducible with your
crush map:
you should open a bug at http://tracker.ceph.com/ to get this fixed.

Deleting the weight set fixes the problem. Moving the item manually with
manual adjustment of the weight set also works in my quick test.

Paul


2018-03-16 16:03 GMT+01:00 <warren.je...@stfc.ac.uk>:

> Hi Paul
>
>
>
> Many thanks for the reply.
>
>
>
> The command is: crush move rack04  room=R80-Upper
>
>
>
> Crush map is here: https://pastebin.com/CX7GKtBy
>
> I’ve done some more testing, and the following all work:
>
> ·         Moving machines between the racks under the default root.
>
> ·         Renaming racks/hosts under the default root
>
> ·         Renaming the default root
>
> ·         Creating a new root
>
> ·         Adding rack05 and rack04 + hosts nina408 and nina508 into the
> new root
>
>
>
> But when trying to move  anything into the default root it fails.
>
>
>
> I have tried moving the following into default root:
>
> ·         Nina408 – with hosts in and without
>
> ·         Nina508 – with hosts in and without
>
> ·         Rack04
>
> ·         Rack05
>
> ·         Rack03 – which I created with nothing in it to try and move.
>
>
>
>
>
> Since first email, I have got the cluster to HEALTH_OK with reweight
> mapping drives, so everything cluster wise appears to be functioning fine.
>
>
>
> I have not tried manually editing the crush map and reimporting for the
> risk that it makes the cluster fall over, as this is currently in
> production. With the CLI I can at least cancel the command the monitor
> comes back up fine.
>
>
>
> Many thanks.
>
>
>
> Warren
>
>
>
>
>
> *From:* Paul Emmerich [mailto:paul.emmer...@croit.io]
> *Sent:* 16 March 2018 13:54
> *To:* Jeffs, Warren (STFC,RAL,ISIS) <warren.je...@stfc.ac.uk>
> *Cc:* ceph-us...@ceph.com
> *Subject:* Re: [ceph-users] Crush Bucket move crashes mons
>
>
>
> Hi,
>
> the error looks like there might be something wrong with the device
> classes (which are managed via separate trees with magic names behind the
> scenes).
>
>
> Can you post your crush map and the command that you are trying to run?
>
> Paul
>
>
>
> 2018-03-15 16:27 GMT+01:00 <warren.je...@stfc.ac.uk>:
>
> Hi All,
>
>
>
> Having some interesting challenges.
>
>
>
> I am trying to move 2 new nodes + 2 new racks into my default root, I have
> added them to the cluster outside of the Root=default.
>
>
>
> They are all in and up – happy it seems. The new nodes have all 12 OSDs in
> them and they are all ‘UP’
>
>
>
> So when going to move them into the correctly room bucket under the
> default root they fail.
>
>
>
> This is the error log at the time: https://pastebin.com/mHfkEp3X
>
>
>
> I can create another host in the crush and move that in and out of rack
> buckets – all while being outside of the default root. Trying to move an
> empty Rack bucket into the default root fails too.
>
>
>
> All of the cluster is on 12.2.4. I do have 2 backfill full osds which is
> the reason for needing these disks in the cluster asap.
>
>
>
> Any thoughts?
>
>
>
> Cheers
>
>
>
> Warren Jeffs
>
>
>
> ISIS Infrastructure Services
>
> STFC Rutherford Appleton Laboratory
>
> e-mail:  warren.je...@stfc.ac.uk
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> --
>
> --
> Paul Emmerich
>
> croit GmbH
> Freseniusstr. 31h
> <https://maps.google.com/?q=Freseniusstr.+31h+%0D%0A81247+M%C3%BCnchen&entry=gmail&source=g>
> 81247 München
> www.croit.io
> Tel: +49 89 1896585 90 <+49%2089%20189658590>
>



-- 
-- 
Paul Emmerich

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to