Frédéric,

When I was writing the last email, my colleague launched a re-peering of the PG in activating state: the PG became active immediately but triggered a little bit of rebalancing of other PGs, not necessarily in the same pool. After this success, we decided to go for your approach, selected a not too critical pool and did a repeer command on all the pool PGs. This resulted in a huge rebalancing (5M objects, in progress, affecting many pools), basically a rebalancing similar (in size) to the unexpected one we have seen after the incident 2 days ago. Could it mean that the state of some OSD was improperly set/used after the restart of OSD servers after the incident and may have resulted in an inappropriate placement of PGs that is being currently fixed after the repeer command causes a reevaluation of the crush map ?

Cheers,

Michel

Le 27/03/2025 à 12:16, Michel Jouvin a écrit :
Frédéric,

Thanks for your answer. I checked the number of PG on osd.17: it is 164, very far from the hard limit (750, the default I think). So it doesn't seem to be the problem and may be the peering is a victim of the more general problem leading to many pools to be more or less inaccessible. What inaccessible means here is not entirely clear:

- We tested the ability to access the pool content with 'rados ls' as I said and we considered that a pool was inaccessible when the command was timing out after 10s (no explicit error). This happens also on empty pools.

- At the same time, on one such pool at least, we were able to successfully upload and download a large file with a S3 client (this pool is part of the data pool of a Swift RGW).

To be honest we have not checked all the logs yet! We concentrated mainly on the mon logs but we'll have a look to some OSD logs.

As for restarting daemons, I am not so reluctant to do it. I have the feeling that in the absence of any message related to inconsistencies, there is no real risk if we restart them one by one and check with ok-to-stop before doing it. What's your feeling? Is it worth restarting the 3 mon first (one by one)?

You mention as an alternative re-peering all PGs of one pool.I was not aware we could do it but I see that there is a 'ceph pg repeer' command. Anything else we should do before running the command? Does it make sense to try it on the PG stucked in activating+remapped state?

Best regards,

Michel

Le 27/03/2025 à 11:40, Frédéric Nass a écrit :
echo "`ceph config get osd.0 mon_max_pg_per_osd`*`ceph config get osd.0 osd_max_pg_per_osd_hard_ratio`" | bc
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to