Hi. Thank you for your reply
The balancer is off, so it shouldn't be interfering with that. Do you happen to have a list of steps (maybe commands from the history) on how you launched it? We're trying to figure out what we're missing.
Regards, Michal On 6/21/25 15:21, Wesley Dillingham wrote:
Michal; Is the balancer (upmap balancer) running (ceph balancer status) and undoing some of the upmaps the pgremapper is introducing? Respectfully, *Wes Dillingham* LinkedIn <http://www.linkedin.com/in/wesleydillingham> w...@wesdillingham.com On Sat, Jun 21, 2025 at 2:05 AM Michal Strnad <michal.str...@cesnet.cz> wrote:Hi. Over the past few days, we've been working on migrating data from machine A to machine B using the pgremapper tool, but we haven’t been able to achieve the expected results. As part of our testing, we set up a small Ceph cluster with several monitors, managers, and servers with OSDs.We applied the flags noout, nobackfill, norecovery, and norebalance, and then added additional servers with OSDs. While Ceph did allocate PG replicas to the newly added OSDs, the actual data didn’t move due to the active flags. We then attempted to use pgremapper to migrate all PGs from one server to the new one, removing or negating the flags in the process. However, we frequently failed to complete the migration of all data/PGs. Are we overlooking something? Does anyone have a reliable, step-by-step procedure we can follow to perform this correctly? Any help would be greatly appreciated! Michal On 3/19/25 08:13, Janne Johansson wrote:The safest approach would be to use the upmap-remapped.py tooldeveloped by Dan at CERN. See [1] for details.The idea is to leverage the upmap load balancer to progressivelymigrate the data to the new servers, minimizing performance impact on the cluster and clients. I like to create the OSDs ahead of time on the nodes that I initially place in a root directory called ‘closet’.I then apply the norebalance flag (ceph osd set norebalance), disablethe balancer (ceph balancer off), move the new nodes with already provisioned OSDs to their final location (rack), run ./upmap-remapped.py to bring all PGs back to active+clean state, remove the norebalance flag (ceph osd unset norebalance), re-enable the balancer (ceph balancer on) and watch data moving progressively as the upmap balancer executes its plans.We do exactly that also, sometimes using pgremapper instead of upmap-remapper.py, but the effect is the same. Make the changes with norebalance, upmap the PGs to be happy where they are until we unset norebalance and let the ceph balancer correct it X% at a time._______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
-- Michal Strnad Storage specialist CESNET a.l.e.
smime.p7s
Description: S/MIME Cryptographic Signature
_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io