Hey Martin, Alright then, we'll just go with the update of every osd's location at once then; just wanted to be sure this was not a problem. :)
On Tue, Oct 22, 2019 at 1:21 PM Martin Verges <martin.ver...@croit.io> wrote: > Hello Alexandre, > > maybe you take a look into https://www.youtube.com/watch?v=V33f7ipw9d4 where > you can see how easy Ceph CRUSH can be managed. > > 1. Changing the locations of all hosts at once >> We are worried that this will generate too much IO and network activity >> (and there is no way to pause / throttle this AFAIK). Maybe this is not >> actually an issue? > > > Just configure the cluster to allow slow recovery before changing the > crush map. Typical options that might help you are "osd recovery sleep > hdd|hybrid|ssd" and "osd max backfills". > > 2. Changing the locations of a couple hosts to reduce data movement >> We are afraid that if we set 2 hosts to DC1, 2 hosts to DC2 and leave the >> rest as-is; Ceph will behave as if there are 3 DCs and will try and fill >> those 4 hosts with as many replicas as possible until they are full. >> > > If you leave any data unsorted, you will never know what data copies are > getting unavailable. In fact, you will produce service impact with such > setups in case of one data center fails. > Do you use any EC configuration suitable for 2 DC configurations, or do > you use replica and want to tolerate having 2 missing copies at the same > time? > > 3. Try and move PGs ahead of the change? >> Maybe we could move PGs so that each PG has a replica on an OSD of each >> DC *before* updating the crush map so that the update does not have to >> actually move any data? (which would allow us to do this at the desired >> pace) >> > > Maybe the PG UPMAP is something that you can use for this, but your > cluster hardware and configuration should always be configured to handle > workloads like this rebalance without impacting your clients. See 1. > > 4. Something else? >> Thank you for your time and your help. :) >> > > You are welcome as every Ceph user! ;) > > -- > Martin Verges > Managing director > > Mobile: +49 174 9335695 > E-Mail: martin.ver...@croit.io > Chat: https://t.me/MartinVerges > > croit GmbH, Freseniusstr. 31h, 81247 Munich > CEO: Martin Verges - VAT-ID: DE310638492 > Com. register: Amtsgericht Munich HRB 231263 > > Web: https://croit.io > YouTube: https://goo.gl/PGE1Bx > > > Am Di., 22. Okt. 2019 um 11:37 Uhr schrieb Alexandre Berthaud < > alexandre.berth...@clever-cloud.com>: > >> Hello everyone, >> >> We have a Ceph cluster (running 14.2.2) which has already dozens of TB of >> data and... we did not set the location of the OSD hosts. The hosts are >> located in 2 datacenters. We would like to update the locations of all the >> hosts so not all replicas end up in a single DC. >> >> We are wondering how we should go about this. >> >> 1. Changing the locations of all hosts at once >> >> We are worried that this will generate too much IO and network activity >> (and there is no way to pause / throttle this AFAIK). Maybe this is not >> actually an issue? >> >> 2. Changing the locations of a couple hosts to reduce data movement >> >> We are afraid that if we set 2 hosts to DC1, 2 hosts to DC2 and leave the >> rest as-is; Ceph will behave as if there are 3 DCs and will try and fill >> those 4 hosts with as many replicas as possible until they are full. >> >> 3. Try and move PGs ahead of the change? >> >> Maybe we could move PGs so that each PG has a replica on an OSD of each >> DC *before* updating the crush map so that the update does not have to >> actually move any data? (which would allow us to do this at the desired >> pace) >> >> 4. Something else? >> >> Thank you for your time and your help. :) >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com