Hi Torkil, Maybe im overlooking something, but how about just renaming the datacenter buckets?
Best regards, Gunnar --- Original Nachricht --- Betreff: [ceph-users] Re: Safe to move misplaced hosts between failure domains in the crush tree? Von: "Torkil Svensgaard" An: "Matthias Grandl" CC: ceph-users@ceph.io, "Ruben Vestergaard" Datum: 12-06-2024 10:33 On 12/06/2024 10:22, Matthias Grandl wrote: > Correct, this should only result in misplaced objects. > > > We made a mistake when we moved the servers physically so while the > replica 3 is intact the crush tree is not accurate. > > Can you elaborate on that? Does this mean after the move, multiple hosts > are inside the same physical datacenter? In that case, once you correct > the CRUSH layout, you would be running misplaced without a way to > rebalance pools that are you using a datacenter crush rule. Hi Matthias Thanks for replying. Two of the three hosts was swapped so I would do: ceph osd crush move ceph-flash1 datacenter=HX1 ceph osd crush move ceph-flash2 datacenter=714 And end up with 2/3 misplaced: -1 4437.29248 root default -33 1467.84814 datacenter 714 -69 69.86389 host ceph-flash2 -34 1511.25378 datacenter HX1 -73 69.86389 host ceph-flash1 -36 1458.19067 datacenter UXH -77 69.86389 host ceph-flash3 It would only briefly be invalid between the two commands. Mvh. Torkil > Cheers! > > -- > > Matthias Grandl > Head Storage Engineer > matthias.gra...@croit.io > > Looking for help with your Ceph cluster? Contact us at https://croit > .io > > croit GmbH, Freseniusstr. 31h, 81247 Munich > CEO: Martin Verges - VAT-ID: DE310638492 > Com. register: Amtsgericht Munich HRB 231263 > Web: https://croit.io | YouTube: https://goo.gl/PGE1Bx > >> On 12. Jun 2024, at 09:13, Torkil Svensgaard wrote: >> >> Hi >> >> We have 3 servers for replica 3 with failure domain datacenter: >> >> -1 4437.29248 root default >> -33 1467.84814 datacenter 714 >> -69 69.86389 host ceph-flash1 >> -34 1511.25378 datacenter HX1 >> -73 69.86389 host ceph-flash2 >> -36 1458.19067 datacenter UXH >> -77 69.86389 host ceph-flash3 >> >> We made a mistake when we moved the servers physically so while the >> replica 3 is intact the crush tree is not accurate. >> >> If we just remedy the situation with "ceph osd crush move ceph-flashX >> datacenter=Y" we will just end up with a lot of misplaced data and >> some churn, right? Or will the affected pool go degraded/unavailable? >> >> Mvh. >> >> Torkil >> -- >> Torkil Svensgaard >> Sysadmin >> MR-Forskningssektionen, afs. 714 >> DRCMR, Danish Research Centre for Magnetic Resonance >> Hvidovre Hospital >> Kettegård Allé 30 >> DK-2650 Hvidovre >> Denmark >> Tel: +45 386 22828 >> E-mail: tor...@drcmr.dk >> _______________________________________________ >> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an email to ceph-users-le...@ceph.io > -- Torkil Svensgaard Sysadmin MR-Forskningssektionen, afs. 714 DRCMR, Danish Research Centre for Magnetic Resonance Hvidovre Hospital Kettegård Allé 30 DK-2650 Hvidovre Denmark Tel: +45 386 22828 E-mail: tor...@drcmr.dk _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io