Thanks Eugen + Frédéric for the hints,currently I am still fishing for reasons - but it seems that the osd has been removed, but the 'purge' did not finish completely.
ceph -s is just doing scrubs: # ceph -s cluster: id: e776dd57-7fc6-11ee-9f23-9bb83aca7b4b health: HEALTH_OK (muted: OSD_UNREACHABLE) services: mon: 5 daemons, quorum aadm01,aadm02,aadm03,aadm05,aadm04 (age 5d)mgr: aadm01.itfgjw(active, since 2d), standbys: aadm02.qlitor, aadm05.nbigji
mds: 2/2 daemons up, 2 standby osd: 1449 osds: 1446 up (since 3d), 1446 in (since 5d) rgw: 2 daemons active (2 hosts, 1 zones) data: volumes: 1/1 healthy pools: 11 pools, 16641 pgs objects: 2.23G objects, 7.0 PiB usage: 8.9 PiB used, 16 PiB / 25 PiB avail pgs: 16604 active+clean 30 active+clean+scrubbing+deep 7 active+clean+scrubbing io: client: 1.8 MiB/s rd, 84 MiB/s wr, 1.68k op/s rd, 2.67k op/s wrThe mgr logs do not show any errors - I just see it trying repeatedly to dispatch the osd removal:
Jun 30 17:39:26 aadm01 bash[1966]: debug 2025-06-30T15:39:26.833+0000 7f1a0e46e640 0 log_channel(audit) log [INF] : from='mgr.108447605 ' entity='mgr.aadm01.itfgjw' cmd=[{"prefix": "osd destroy-actual", "id": 406, "yes_i_really_mean_it": true}]: dispatch
Looking at the cephadm logs I see:2025-06-30T15:39:26.836933+0000 mgr.aadm01.itfgjw (mgr.108447605) 368120 : cephadm [INF] Daemon osd.406 on acn07 was already removed 2025-06-30T15:39:26.839850+0000 mgr.aadm01.itfgjw (mgr.108447605) 368121 : cephadm [INF] Successfully destroyed old osd.406 on acn07; ready for replacement
And ater stopping the removal and trying to reissue it I get: # ceph orch osd rm stop 406 Stopped OSD(s) removal # ceph orch osd rm 406 --replace --force Unable to find OSDs: ['406'] So it seems that it was finished, but somehow did not finish cleanly. An 'ceph osd status' shows406 0 0 0 0 0 0 destroyed,exists
for osd.406.I will try to continue with the disk replacement and see where and if any more 'manual' intervention is needed for this.
Regards, Holger On 30.06.25 11:32, Frédéric Nass wrote:
Hi Holger, In addition to Eugen's sound advice, I would try restarting the OSD in question. If that doesn't help, I would stop the current deletion process 'ceph orch osd rm stop 406' and restart it 'ceph orch osd rm 406 --force'. As for the orchestrator logs, you can use the command 'ceph log last 1000 debug cephadm' and also check for any errors in ceph log files on host 'acn07'. Regards, Frédéric. ----- Le 28 Juin 25, à 19:53, Eugen Block ebl...@nde.ag a écrit :Can you show the overall cluster status (ceph -s)? If there's something else going on, it might block (some?) operations. And I'd scan the mgr logs, maybe in debug mode to see why it fails to operate properly. Zitat von Holger Naundorf <naund...@rz.uni-kiel.de>:On 27.06.25 14:16, Eugen Block wrote:Zitat von Holger Naundorf <naund...@rz.uni-kiel.de>:Hello, title should of course be "orchestrator behaving strangely" I did give a mgr restart another try (for the last OSD removal - which also did not work - I did already restart the mgr without effect) there is no (immediate- i.e after ~10min) effect now as well - or should I reissue the OSD rm command as well?Is there something in the queue (ceph orch osd rm status)? Sometimes the queue clears after a mgr restart, so it might be necessary to restart the rm command as well.There is just the one 'waiting for purge' osd in the queue: root@aadm01:~# ceph orch osd rm status OSD HOST STATE PGS REPLACE FORCE ZAP DRAIN STARTED AT 406 acn07 done, waiting for purge 0 True False True 2025-06-25 09:18:07.650734+00:00 One more datapoint: This is an OSD on a lage, rotational disk. The orchestrator is still working ok for a subet of OSDs on SSDs. We are just moving our SSD pool around and for this it was no problem using 'ceph orch osd rm ...' - with the difference that there we used --zap and not --replace (as we do not want to replace the disk, but to move the OSD away from this host). Regards, HolgerRegards, Holger On 27.06.25 12:26, Eugen Block wrote:Hi, have you retried it after restarting/failing the mgr? ceph mgr fail Quite often this (still) helps. Zitat von Holger Naundorf <naund...@rz.uni-kiel.de>:Hello, we are running a ceph cluster at version: ceph version 19.2.2 (0eceb0defba60152a8182f7bd87d164b639885b8) squid (stable) and since a few weeks the orchestrator started to misbehave - up to now we could not identify any root cause, so I am fishing in the community to see if there are any hints. Problems: An OSD removal (for disk replacement) gets stuck in the 'purge' step: ceph orch osd rm 406 --replace root@aadm01:~# ceph orch osd rm status OSD HOST STATE PGS REPLACE FORCE ZAP DRAIN STARTED AT 406 acn07 done, waiting for purge 0 True False True 2025-06-25 09:18:07.650734+00:00 (now for more than 24h in this state) At the same time the orchestrator is not restarting OSD daemons - i.e. an 'ceph orch daemon restart osd.xxx' claims its queuing uo the restart, but it never happens. Other services continue to be controlled correctly via 'ceph orch ...' If anyone has an idea where to poke around or can match this to some known problem - I would appreciate any pointers. Regards, Holger -- Dr. Holger Naundorf Christian-Albrechts-Universität zu Kiel Rechenzentrum / HPC / Server und Storage Tel: +49 431 880-1990 Fax: +49 431 880-1523 naund...@rz.uni-kiel.de_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io-- Dr. Holger Naundorf Christian-Albrechts-Universität zu Kiel Rechenzentrum / HPC / Server und Storage Tel: +49 431 880-1990 Fax: +49 431 880-1523 naund...@rz.uni-kiel.de_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io-- Dr. Holger Naundorf Christian-Albrechts-Universität zu Kiel Rechenzentrum / HPC / Server und Storage Tel: +49 431 880-1990 Fax: +49 431 880-1523 naund...@rz.uni-kiel.de_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
-- Dr. Holger Naundorf Christian-Albrechts-Universität zu Kiel Rechenzentrum / HPC / Server und Storage Tel: +49 431 880-1990 Fax: +49 431 880-1523 naund...@rz.uni-kiel.de
smime.p7s
Description: S/MIME Cryptographic Signature
_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io