I usually add 20 OSDs each time.
To take control of the influence of backfilling, I will set
primary-affinity to 0 of those new OSDs and adjust backfilling
configurations.
http://docs.ceph.com/docs/master/rados/configuration/osd-config-ref/#backfilling


Kevin Hrpcek <kevin.hrp...@ssec.wisc.edu> 于2019年7月25日周四 上午2:02写道:

> I change the crush weights. My 4 second sleep doesn't let peering finish
> for each one before continuing. I'd test with some small steps to get an
> idea of how much remaps when increasing the weight by $x. I've found my
> cluster is comfortable with +1 increases...also it take awhile to get to a
> weight of 11 if I did anything smaller.
>
> for i in {264..311}; do ceph osd crush reweight osd.${i} 11.0;sleep 4;done
>
> Kevin
>
> On 7/24/19 12:33 PM, Xavier Trilla wrote:
>
> Hi Kevin,
>
> Yeah, that makes a lot of sense, and looks even safer than adding OSDs one
> by one. What do you change, the crush weight? Or the reweight? (I guess you
> change the crush weight, I am right?)
>
> Thanks!
>
>
>
> El 24 jul 2019, a les 19:17, Kevin Hrpcek <kevin.hrp...@ssec.wisc.edu> va
> escriure:
>
> I often add 50+ OSDs at a time and my cluster is all NLSAS. Here is what I
> do, you can obviously change the weight increase steps to what you are
> comfortable with. This has worked well for me and my workloads. I've
> sometimes seen peering take longer if I do steps too quickly but I don't
> run any mission critical has to be up 100% stuff and I usually don't notice
> if a pg takes a while to peer.
>
> Add all OSDs with an initial weight of 0. (nothing gets remapped)
> Ensure cluster is healthy.
> Use a for loop to increase weight on all news OSDs to 0.5 with a generous
> sleep between each for peering.
> Let the cluster balance and get healthy or close to healthy.
> Then repeat the previous 2 steps increasing weight by +0.5 or +1.0 until I
> am at the desired weight.
>
> Kevin
>
> On 7/24/19 11:44 AM, Xavier Trilla wrote:
>
> Hi,
>
>
>
> What would be the proper way to add 100 new OSDs to a cluster?
>
>
>
> I have to add 100 new OSDs to our actual > 300 OSDs cluster, and I would
> like to know how you do it.
>
>
>
> Usually, we add them quite slowly. Our cluster is a pure SSD/NVMe one, and
> it can handle plenty of load, but for the sake of safety -it hosts
> thousands of VMs via RBD- we usually add them one by one, waiting for a
> long time between adding each OSD.
>
>
>
> Obviously this leads to PLENTY of data movement, as each time the cluster
> geometry changes, data is migrated among all the OSDs. But with the kind of
> load we have, if we add several OSDs at the same time, some PGs can get
> stuck for a while, while they peer to the new OSDs.
>
>
>
> Now that I have to add > 100 new OSDs I was wondering if somebody has some
> suggestions.
>
>
>
> Thanks!
>
> Xavier.
>
> _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to