The balancer is the usual way of using a remapper.

* Set norebalance
* Create the OSDs
* Run the remapper so the backfill goes away
* Unset norebalance
* The balancer then will incrementally remove the extra upmaps, gradually 
evening out data.


What did you do?  Did you compute a set of upmaps by hand and apply them?

> On Jun 21, 2025, at 11:00 AM, Michal Strnad <michal.str...@cesnet.cz> wrote:
> 
> Hi.
> 
> Thank you for your reply
> 
> The balancer is off, so it shouldn't be interfering with that. Do you happen 
> to have a list of steps (maybe commands from the history) on how you launched 
> it? We're trying to figure out what we're missing.
> 
> Regards,
> Michal
> 
> 
> On 6/21/25 15:21, Wesley Dillingham wrote:
>> Michal;
>> Is the balancer (upmap balancer) running (ceph balancer status) and undoing
>> some of the upmaps the pgremapper is introducing?
>> Respectfully,
>> *Wes Dillingham*
>> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>> w...@wesdillingham.com
>> On Sat, Jun 21, 2025 at 2:05 AM Michal Strnad <michal.str...@cesnet.cz>
>> wrote:
>>> Hi.
>>> 
>>> Over the past few days, we've been working on migrating data from
>>> machine A to machine B using the pgremapper tool, but we haven’t been
>>> able to achieve the expected results.
>>> 
>>> As part of our testing, we set up a small Ceph cluster with several
>>> monitors, managers, and servers with OSDs.We applied the flags noout,
>>> nobackfill, norecovery, and norebalance, and then added additional
>>> servers with OSDs. While Ceph did allocate PG replicas to the newly
>>> added OSDs, the actual data didn’t move due to the active flags. We then
>>> attempted to use pgremapper to migrate all PGs from one server to the
>>> new one, removing or negating the flags in the process. However, we
>>> frequently failed to complete the migration of all data/PGs.
>>> Are we overlooking something? Does anyone have a reliable, step-by-step
>>> procedure we can follow to perform this correctly?
>>> 
>>> Any help would be greatly appreciated!
>>> 
>>> Michal
>>> 
>>> 
>>> On 3/19/25 08:13, Janne Johansson wrote:
>>>>> The safest approach would be to use the upmap-remapped.py tool
>>> developed by Dan at CERN. See [1] for details.
>>>>> 
>>>>> The idea is to leverage the upmap load balancer to progressively
>>> migrate the data to the new servers, minimizing performance impact on the
>>> cluster and clients. I like to create the OSDs ahead of time on the nodes
>>> that I initially place in a root directory called ‘closet’.
>>>>> 
>>>>> I then apply the norebalance flag (ceph osd set norebalance), disable
>>> the balancer (ceph balancer off), move the new nodes with already
>>> provisioned OSDs to their final location (rack), run ./upmap-remapped.py to
>>> bring all PGs back to active+clean state, remove the norebalance flag (ceph
>>> osd unset norebalance), re-enable the balancer (ceph balancer on) and watch
>>> data moving progressively as the upmap balancer executes its plans.
>>>> 
>>>> We do exactly that also, sometimes using pgremapper instead of
>>>> upmap-remapper.py, but the effect is the same. Make the changes with
>>>> norebalance, upmap the PGs to be happy where they are until we unset
>>>> norebalance and let the ceph balancer correct it X% at a time.
>>>> 
>>> 
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@ceph.io
>>> To unsubscribe send an email to ceph-users-le...@ceph.io
>>> 
> 
> -- 
> Michal Strnad
> Storage specialist
> CESNET a.l.e.
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to