[ceph-users] Re: Accidentally created systemd units for OSDs

2024-08-16 Thread Dan O'Brien
OK... I've been in the Circle of Hell where systemd lives and I *THINK* I have convinced myself I'm OK. I *REALLY* don't want to trash and rebuild the OSDs. In the manpage for systemd.unit, I found UNIT GARBAGE COLLECTION The system and service manager loads a unit's configuration automatically

[ceph-users] Re: Accidentally created systemd units for OSDs

2024-08-16 Thread Dan O'Brien
I am 100% using cephadm and containers and plan to continue to do so. Our original setup was all spinners, but after going to Ceph Days NYC, I pushed for SSDs to use for the WAL/RocksDb and I'm in the process of migrating the WAL/RocksDb. In general, it's been fairly straightforward -- IF YOU FO

[ceph-users] Accidentally created systemd units for OSDs

2024-08-16 Thread Dan O'Brien
I was [poorly] following the instructions for migrating the wal/db to an SSD https://docs.clyso.com/blog/ceph-volume-create-wal-db-on-separate-device-for-existing-osd/ and I didn't add the '--no-systemd' when I did 'ceph-volume lvm activate' command (3 f***ing times). The result is that I've "tw

[ceph-users] Re: Dashboard error on 18.2.4 when listing block images

2024-07-26 Thread Dan O'Brien
I'll try that today. Looking at the tracker issue you flagged, it seems like it should be fixed in v18.2.4, which is what I'm running. Did that commit make it into the 18.2.4 build that was released? ___ ceph-users mailing list -- ceph-users@ceph.io To

[ceph-users] Re: Dashboard error on 18.2.4 when listing block images

2024-07-25 Thread Dan O'Brien
Ilya - I don't think images-pubos/144ebab3-b2ee-4331-9d41-8505bcc4e19b is the problem; it was just the last RBD image listed in the log before the crash. The commands you suggested work fine when using that image: [root@os-storage ~]# rbd info images-pubos/144ebab3-b2ee-4331-9d41-8505bcc4e19b r

[ceph-users] Re: Dashboard error on 18.2.4 when listing block images

2024-07-25 Thread Dan O'Brien
Both of the volumes mentioned above are showing a status of "error_deleting" in OpenStack. I'm probably going to have to remove them with openstack volume delete --force. I'm guessing this might be a regression in DiffIterate that doesn't handle volumes/RBD images in an inconsistent state. It

[ceph-users] Re: Dashboard error on 18.2.4 when listing block images

2024-07-25 Thread Dan O'Brien
I may be on to something. Going through the pools and listing the block images in them, I have one pool that gives an error message: [root@os-storage ~]# rbd --pool volumes-gpu ls --long rbd: error opening volume-28bbca8c-fec5-4a33-bbe2-30408f1ea37f: (2) No such file or directory rbd: error open

[ceph-users] Dashboard error on 18.2.4 when listing block images

2024-07-25 Thread Dan O'Brien
I just yesterday updated by Ceph cluster from Quincy (17.2.7) to Reef (18.2.4, purely by accident; I didn't realize I was getting it THE EXACT SECOND it was pushed out). The upgrade resolved an issue I was having with HTTP 500 errors on the RGW UI, but seems to have created an issue in the dash

[ceph-users] Re: Subscribe

2024-07-25 Thread Dan O'Brien
Sorry, the list has been wonky for me. I was logged in with my GitHub credentials and when I try and publish the post, I get the message This list is moderated, please subscribe to it before posting. When I try and manage my subscription, I get: Something went wrong Mailman REST API not avai

[ceph-users] Subscribe

2024-07-25 Thread Dan O'Brien
___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] OSD containers lose connectivity after change from Rocky 8.7->9.2

2023-08-15 Thread Dan O'Brien
I recently updated one of the hosts (an older Dell PowerEdge R515) in my Ceph Quincy (17.2.6) cluster. I needed to change the IP address, so I removed the host from the cluster (gracefully removed OSDs and daemons, then removed the host). I also took the opportunity to upgrade the host from Rock