Michel, I can't recall any situations like that - maybe someone here does? - but I would advise that you restart all OSDs to trigger the re-peering of every PG. This should get your cluster back on track.
Just make sure the crush map / crush rules / bucket weights (including OSDs weights) haven't changed, as this would of course trigger rebalancing. Regards, Frédéric. ----- Le 27 Mar 25, à 13:30, Michel Jouvin michel.jou...@ijclab.in2p3.fr a écrit : > Frédéric, > > When I was writing the last email, my colleague launched a re-peering of > the PG in activating state: the PG became active immediately but > triggered a little bit of rebalancing of other PGs, not necessarily in > the same pool. After this success, we decided to go for your approach, > selected a not too critical pool and did a repeer command on all the > pool PGs. This resulted in a huge rebalancing (5M objects, in progress, > affecting many pools), basically a rebalancing similar (in size) to the > unexpected one we have seen after the incident 2 days ago. Could it mean > that the state of some OSD was improperly set/used after the restart of > OSD servers after the incident and may have resulted in an inappropriate > placement of PGs that is being currently fixed after the repeer command > causes a reevaluation of the crush map ? > > Cheers, > > Michel > > Le 27/03/2025 à 12:16, Michel Jouvin a écrit : >> Frédéric, >> >> Thanks for your answer. I checked the number of PG on osd.17: it is >> 164, very far from the hard limit (750, the default I think). So it >> doesn't seem to be the problem and may be the peering is a victim of >> the more general problem leading to many pools to be more or less >> inaccessible. What inaccessible means here is not entirely clear: >> >> - We tested the ability to access the pool content with 'rados ls' as >> I said and we considered that a pool was inaccessible when the command >> was timing out after 10s (no explicit error). This happens also on >> empty pools. >> >> - At the same time, on one such pool at least, we were able to >> successfully upload and download a large file with a S3 client (this >> pool is part of the data pool of a Swift RGW). >> >> To be honest we have not checked all the logs yet! We concentrated >> mainly on the mon logs but we'll have a look to some OSD logs. >> >> As for restarting daemons, I am not so reluctant to do it. I have the >> feeling that in the absence of any message related to inconsistencies, >> there is no real risk if we restart them one by one and check with >> ok-to-stop before doing it. What's your feeling? Is it worth >> restarting the 3 mon first (one by one)? >> >> You mention as an alternative re-peering all PGs of one pool.I was not >> aware we could do it but I see that there is a 'ceph pg repeer' >> command. Anything else we should do before running the command? Does >> it make sense to try it on the PG stucked in activating+remapped state? >> >> Best regards, >> >> Michel >> >> Le 27/03/2025 à 11:40, Frédéric Nass a écrit : >>> echo "`ceph config get osd.0 mon_max_pg_per_osd`*`ceph config get > >> osd.0 osd_max_pg_per_osd_hard_ratio`" | bc _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io