(Adding back the list)

https://docs.clyso.com/docs/kb/rados/#osd---improved-procedure-for-adding-hosts-or-osds
 
<https://docs.clyso.com/docs/kb/rados/#osd---improved-procedure-for-adding-hosts-or-osds>

Is the upstream source for `upmap-remapped.py` the `ceph-scripts` repo?

If yes, is it known why there are 2 different places for `ceph-scripts` (Gitlab 
and Github), with diverged histories, both of which are recently edited, by the 
same individuals?

https://github.com/cernceph/ceph-scripts/commits/master/tools/upmap/upmap-remapped.py
https://gitlab.cern.ch/ceph/ceph-scripts/-/commits/master/tools/upmap/upmap-remapped.py

If I understand what you’re getting at, I don’t think that’s the case.  I 
*think* recent releases give priority to undersized / degraded PGs over 
remapped PGs.

It's not really `degraded` vs `misplaced` I'm getting at.

But instead, consider this scneario:

We have

* osd.1 (very full, 100 GB free)
* osd.2 (very full, 100 GB free)
* osd.30 (empty)

and we have

* PGa on osd.1, which Ceph now wants to be on osd.2
* PGb on osd.2, which Ceph now wants to be on osd.30

The simple thing that could theoretically be done is that simultaneously be done is 
that the PGa (osd.1 -> osd.2) move, and the PGb (osd2. -> osd.30) move, are 
started immediately and run in parallel.

But I suspect that what Ceph does is to make decisions on PG granularity:

1. See that there's no space on osd.2 for PGa, thus not start the move.
2. Move PGb from osd.2 to osd.30. Wait until that's done (takes long if PGs are 
large).
3. See that now there enough space on osd.2 for PGa, thus only now start the 
move.

Is this really what happens?

Thanks!
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to