(Adding back the list)
https://docs.clyso.com/docs/kb/rados/#osd---improved-procedure-for-adding-hosts-or-osds <https://docs.clyso.com/docs/kb/rados/#osd---improved-procedure-for-adding-hosts-or-osds>
Is the upstream source for `upmap-remapped.py` the `ceph-scripts` repo? If yes, is it known why there are 2 different places for `ceph-scripts` (Gitlab and Github), with diverged histories, both of which are recently edited, by the same individuals? https://github.com/cernceph/ceph-scripts/commits/master/tools/upmap/upmap-remapped.py https://gitlab.cern.ch/ceph/ceph-scripts/-/commits/master/tools/upmap/upmap-remapped.py
If I understand what you’re getting at, I don’t think that’s the case. I *think* recent releases give priority to undersized / degraded PGs over remapped PGs.
It's not really `degraded` vs `misplaced` I'm getting at. But instead, consider this scneario: We have * osd.1 (very full, 100 GB free) * osd.2 (very full, 100 GB free) * osd.30 (empty) and we have * PGa on osd.1, which Ceph now wants to be on osd.2 * PGb on osd.2, which Ceph now wants to be on osd.30 The simple thing that could theoretically be done is that simultaneously be done is that the PGa (osd.1 -> osd.2) move, and the PGb (osd2. -> osd.30) move, are started immediately and run in parallel. But I suspect that what Ceph does is to make decisions on PG granularity: 1. See that there's no space on osd.2 for PGa, thus not start the move. 2. Move PGb from osd.2 to osd.30. Wait until that's done (takes long if PGs are large). 3. See that now there enough space on osd.2 for PGa, thus only now start the move. Is this really what happens? Thanks! _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io