Nope, it was really broken in 17.2.7. When RHEL 10 comes available, I will look into this part again :)
> Op 25-04-2025 07:22 CEST schreef Lukasz Borek <luk...@borek.org.pl>: > > > > For upgrade the OS we have something similar, but exiting maintenance mode > > is broken (with 17.2.7) :( > > I need to check the tracker for similar issues and if I can't find > > anything, I will create a ticket > For 18.2.2 first maint exit command threw an exception for some reason. In my > patching script I execute commands in a loop and the 2nd shoot usually works. > > exit maint 1/3 > Error EINVAL: Traceback (most recent call last): > File "/usr/share/ceph/mgr/mgr_module.py", line 1809, in _handle_command > return self.handle_command(inbuf, cmd) > File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 183, in > handle_command > return dispatch[cmd['prefix']].call(self, cmd, inbuf) > File "/usr/share/ceph/mgr/mgr_module.py", line 474, in call > return self.func(mgr, **kwargs) > File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 119, in <lambda> > wrapper_copy = lambda *l_args, **l_kwargs: wrapper(*l_args, **l_kwargs) # > noqa: E731 > File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 108, in wrapper > return func(*args, **kwargs) > File "/usr/share/ceph/mgr/orchestrator/module.py", line 778, in > _host_maintenance_exit > raise_if_exception(completion) > File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 237, in > raise_if_exception > e = pickle.loads(c.serialized_exception) > TypeError: __init__() missing 2 required positional arguments: 'hostname' and > 'addr' > > exit maint 2/3 > Ceph cluster f3e63d9e-2f4c-11ef-87a2-0f1170f55ed5 on cephbackup-osd1 has > exited maintenance mode > exit maint 3/3 > Error EINVAL: Host cephbackup-osd1 is not in maintenance mode > Fri Apr 25 07:17:58 CEST 2025 cluster state is HEALTH_WARN > Fri Apr 25 07:18:02 CEST 2025 cluster state is HEALTH_WARN > [...] > > > > > > On Thu, 13 Jun 2024 at 22:07, Sake Ceph <c...@paulusma.eu> wrote: > > > > > > For upgrade the OS we have something similar, but exiting maintenance mode > > is broken (with 17.2.7) :( > > I need to check the tracker for similar issues and if I can't find > > anything, I will create a ticket. > > > > Kind regards, > > Sake > > > > > Op 12-06-2024 19:02 CEST schreef Daniel Brown > > <daniel.h.brown@thermify.cloud>: > > > > > > > > > I have two ansible roles, one for enter, one for exit. There’s likely > > better ways to do this — and I’ll not be surprised if someone here lets me > > know. They’re using orch commands via the cephadm shell. I’m using Ansible > > for other configuration management in my environment, as well, including > > setting up clients of the ceph cluster. > > > > > > > > > Below excerpts from main.yml in the “tasks” for the enter/exit roles. > > The host I’m running ansible from is one of my CEPH servers - I’ve limited > > which process run there though so it’s in the cluster but not equal to the > > others. > > > > > > > > > ————— > > > Enter > > > ————— > > > > > > - name: Ceph Maintenance Mode Enter > > > shell: > > > > > > cmd: ' cephadm shell ceph orch host maintenance enter {{ > > (ansible_ssh_host|default(ansible_host))|default(inventory_hostname) }} > > --force --yes-i-really-mean-it ‘ > > > become: True > > > > > > > > > > > > ————— > > > Exit > > > ————— > > > > > > > > > - name: Ceph Maintenance Mode Exit > > > shell: > > > cmd: 'cephadm shell ceph orch host maintenance exit {{ > > (ansible_ssh_host|default(ansible_host))|default(inventory_hostname) }} ‘ > > > become: True > > > connection: local > > > > > > > > > - name: Wait for Ceph to be available > > > ansible.builtin.wait_for: > > > delay: 60 > > > host: '{{ > > (ansible_ssh_host|default(ansible_host))|default(inventory_hostname) }}’ > > > port: 9100 > > > connection: local > > > > > > > > > > > > > > > > > > > > > > On Jun 12, 2024, at 11:28 AM, Michael Worsham > > <mwors...@datadimensions.com> wrote: > > > > > > > > Interesting. How do you set this "maintenance mode"? If you have a > > series of documented steps that you have to do and could provide as an > > example, that would be beneficial for my efforts. > > > > > > > > We are in the process of standing up both a dev-test environment > > consisting of 3 Ceph servers (strictly for testing purposes) and a new > > production environment consisting of 20+ Ceph servers. > > > > > > > > We are using Ubuntu 22.04. > > > > > > > > -- Michael > > > > From: Daniel Brown <daniel.h.brown@thermify.cloud> > > > > Sent: Wednesday, June 12, 2024 9:18 AM > > > > To: Anthony D'Atri <anthony.da...@gmail.com> > > > > Cc: Michael Worsham <mwors...@datadimensions.com>; ceph-users@ceph.io > > <ceph-users@ceph.io> > > > > Subject: Re: [ceph-users] Patching Ceph cluster > > > > This is an external email. Please take care when clicking links or > > opening attachments. When in doubt, check with the Help Desk or Security. > > > > > > > > > > > > There’s also a Maintenance mode that you can set for each server, as > > you’re doing updates, so that the cluster doesn’t try to move data from > > affected OSD’s, while the server being updated is offline or down. I’ve > > worked some on automating this with Ansible, but have found my process > > (and/or my cluster) still requires some manual intervention while it’s > > running to get things done cleanly. > > > > > > > > > > > > > > > > > On Jun 12, 2024, at 8:49 AM, Anthony D'Atri > > <anthony.da...@gmail.com> wrote: > > > > > > > > > > Do you mean patching the OS? > > > > > > > > > > If so, easy -- one node at a time, then after it comes back up, wait > > until all PGs are active+clean and the mon quorum is complete before > > proceeding. > > > > > > > > > > > > > > > > > > > >> On Jun 12, 2024, at 07:56, Michael Worsham > > <mwors...@datadimensions.com> wrote: > > > > >> > > > > >> What is the proper way to patch a Ceph cluster and reboot the > > servers in said cluster if a reboot is necessary for said updates? And is > > it possible to automate it via Ansible? This message and its attachments > > are from Data Dimensions and are intended only for the use of the > > individual or entity to which it is addressed, and may contain information > > that is privileged, confidential, and exempt from disclosure under > > applicable law. If the reader of this message is not the intended > > recipient, or the employee or agent responsible for delivering the message > > to the intended recipient, you are hereby notified that any dissemination, > > distribution, or copying of this communication is strictly prohibited. If > > you have received this communication in error, please notify the sender > > immediately and permanently delete the original email and destroy any > > copies or printouts of this email as well as any attachments. > > > > >> _______________________________________________ > > > > >> ceph-users mailing list -- ceph-users@ceph.io > > > > >> To unsubscribe send an email to ceph-users-le...@ceph.io > > > > > _______________________________________________ > > > > > ceph-users mailing list -- ceph-users@ceph.io > > > > > To unsubscribe send an email to ceph-users-le...@ceph.io > > > > > > > > This message and its attachments are from Data Dimensions and are > > intended only for the use of the individual or entity to which it is > > addressed, and may contain information that is privileged, confidential, > > and exempt from disclosure under applicable law. If the reader of this > > message is not the intended recipient, or the employee or agent responsible > > for delivering the message to the intended recipient, you are hereby > > notified that any dissemination, distribution, or copying of this > > communication is strictly prohibited. If you have received this > > communication in error, please notify the sender immediately and > > permanently delete the original email and destroy any copies or printouts > > of this email as well as any attachments. > > > > > > _______________________________________________ > > > ceph-users mailing list -- ceph-users@ceph.io > > > To unsubscribe send an email to ceph-users-le...@ceph.io > > _______________________________________________ > > ceph-users mailing list -- ceph-users@ceph.io > > To unsubscribe send an email to ceph-users-le...@ceph.io > > > > > -- > > Łukasz Borek > luk...@borek.org.pl _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io