> (Adding back the list)

I had written privately on purpose.

> 
>> https://docs.clyso.com/docs/kb/rados/#osd---improved-procedure-for-adding-hosts-or-osds
>>  
>> <https://docs.clyso.com/docs/kb/rados/#osd---improved-procedure-for-adding-hosts-or-osds>
> 
> Is the upstream source for `upmap-remapped.py` the `ceph-scripts` repo?
> 
> If yes, is it known why there are 2 different places for `ceph-scripts` 
> (Gitlab and Github), with diverged histories, both of which are recently 
> edited, by the same individuals?
> 
> https://github.com/cernceph/ceph-scripts/commits/master/tools/upmap/upmap-remapped.py
> https://gitlab.cern.ch/ceph/ceph-scripts/-/commits/master/tools/upmap/upmap-remapped.py

I will defer to Dan van der Ster on that.

> 
>> If I understand what you’re getting at, I don’t think that’s the case.  I 
>> *think* recent releases give priority to undersized / degraded PGs over 
>> remapped PGs.
> 
> It's not really `degraded` vs `misplaced` I'm getting at.
> 
> But instead, consider this scneario:
> 
> We have
> 
> * osd.1 (very full, 100 GB free)
> * osd.2 (very full, 100 GB free)
> * osd.30 (empty)
> 
> and we have
> 
> * PGa on osd.1, which Ceph now wants to be on osd.2
> * PGb on osd.2, which Ceph now wants to be on osd.30
> 
> The simple thing that could theoretically be done is that simultaneously be 
> done is that the PGa (osd.1 -> osd.2) move, and the PGb (osd2. -> osd.30) 
> move, are started immediately and run in parallel.
> 
> But I suspect that what Ceph does is to make decisions on PG granularity:
> 
> 1. See that there's no space on osd.2 for PGa, thus not start the move.
> 2. Move PGb from osd.2 to osd.30. Wait until that's done (takes long if PGs 
> are large).
> 3. See that now there enough space on osd.2 for PGa, thus only now start the 
> move.
> 
> Is this really what happens?

Yes, backfill will not proceed to an OSD that is past the backfillful threshold.
It’s also the case that the cluster usually cannot sustain all backfill / 
recovery ops at the same time; there are priority mechanisms as well as various 
throttles to avoid a thundering herd DoSing client ops.  I will have to defer 
to the source and devs for the particulars of that process, as it has evolved 
over time.

> 
> Thanks!
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to