Hi there!

Here we have a similar situation.
After adding some OSDs to the cluster the PGs are not equally distributed over 
the OSDs.

The balancing mode is set to upmap.
The docs https://docs.ceph.com/docs/master/rados/operations/balancer/#modes say:
"This CRUSH mode will optimize the placement of individual PGs in order to 
achieve a balanced distribution. In most cases, this distribution is “perfect,” 
which an equal number of PGs on each OSD (+/-1 PG, since they might not divide 
evenly)."

This is not the case here with our cluster. The number of PGs reaches from 157 
up to 214. And so the resulting usage of the HDDs varies from 60% to 82%.
$ ceph osd df class hdd
[…]
MIN/MAX VAR: 0.89/1.22  STDDEV: 7.47

In the meantime I had to use reweight-by-utilization to get the cluster in 
healthy state again, because it had a near full PG and near full OSDs.

The cluster is nautilus 14.2.4 on debian 10 with packages from croit.io.

I think about switching back to crush-compat.

Is there a bug known regarding this?

Thanks
Lars
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to