Note that 'norebalance' disables the balancer but doesn't prevent
backfill; you'll want to set 'nobackfill' as well.

Josh

On Sun, Jan 12, 2025 at 1:49 PM Anthony D'Atri <anthony.da...@gmail.com> wrote:
>
> [ ed: snag during moderation (somehow a newline was interpolated in the 
> Subject), so I’m sending this on behalf of kasper_steenga...@hotmail.com 
> <mailto:kasper_steenga...@hotmail.com> , to whom replies should be sent]
>
>
> I'm managing a ceph cluster with +1K OSDs distributed accross 56 host.
> Until now the crush rule used is the default replicated rule, but I want to 
> change that in order to implement failure domain on rack level.
>
> Current plan is to
> - Disable rebalancing by executing - ceph osd set norebalance;
> - Add Rack to crushmap and distribute the hosts accordingly (8 in each.) by 
> using the built in commands
>       - ceph osd crush add-bucket rack1 rack root=default
>       - ceph osd crush move osd-host1 rack=rack1
> - Create the new rack split rule with command
>        -  ceph osd crush rule create-replicated rack_split default rack
> - Set the rule across all my pools
>       - for p in $(ceph osd lspools | cut -d' ' -f 2) ; do echo $p $(ceph osd 
> pool set $p crush_rule rack_split) ; done
> - Finally enable rebalancing - ceph osd unset norebalance;
>
> However I'm concerned with the amount of data that needs to be rebalanced, 
> since
>  the cluster holds multiple PB, and I'm looking for review of/input for my 
> plan,
>  as well as words of advice/experience from someone who have been in similar 
> situations.
>
>
> ——
>
> [
>  ed: You only include commands for one CRUSH `rack` — would you create 
> multiple `rack` CRUSH buckets, at least three of them?
>
> Are all of your pools replicated?  No EC pools for RGW buckets, CephFS data, 
> etc?
>
> What OSD media and networking does this cluster have? HDDs will be much 
> slower and much more impacted during the process than SSDs. Is your client 
> workload 24x7? Which Ceph release? These factors inform how impactful the 
> grand shuffle will be.  Are your mon DBs on SSDs?
>
> A popular strategy is to use upmap-remapped.py to freeze all of the PG 
> mappings before unsetting the norebalance flag, then the balancer will 
> gradually undo the mappings as it moves data to where it now belongs.  This 
> process has built-in throttling.
>
> ]
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to