OK... I've been in the Circle of Hell where systemd lives and I *THINK* I have
convinced myself I'm OK. I *REALLY* don't want to trash and rebuild the OSDs.
In the manpage for systemd.unit, I found
UNIT GARBAGE COLLECTION
The system and service manager loads a unit's configuration automatically
I am 100% using cephadm and containers and plan to continue to do so.
Our original setup was all spinners, but after going to Ceph Days NYC, I pushed
for SSDs to use for the WAL/RocksDb and I'm in the process of migrating the
WAL/RocksDb. In general, it's been fairly straightforward -- IF YOU FO
I was [poorly] following the instructions for migrating the wal/db to an SSD
https://docs.clyso.com/blog/ceph-volume-create-wal-db-on-separate-device-for-existing-osd/
and I didn't add the '--no-systemd' when I did 'ceph-volume lvm activate'
command (3 f***ing times). The result is that I've "tw
I'll try that today.
Looking at the tracker issue you flagged, it seems like it should be fixed in
v18.2.4, which is what I'm running. Did that commit make it into the 18.2.4
build that was released?
___
ceph-users mailing list -- ceph-users@ceph.io
To
Ilya -
I don't think images-pubos/144ebab3-b2ee-4331-9d41-8505bcc4e19b is the problem;
it was just the last RBD image listed in the log before the crash. The commands
you suggested work fine when using that image:
[root@os-storage ~]# rbd info images-pubos/144ebab3-b2ee-4331-9d41-8505bcc4e19b
r
Both of the volumes mentioned above are showing a status of "error_deleting" in
OpenStack. I'm probably going to have to remove them with openstack volume
delete --force.
I'm guessing this might be a regression in DiffIterate that doesn't handle
volumes/RBD images in an inconsistent state.
It
I may be on to something. Going through the pools and listing the block images
in them, I have one pool that gives an error message:
[root@os-storage ~]# rbd --pool volumes-gpu ls --long
rbd: error opening volume-28bbca8c-fec5-4a33-bbe2-30408f1ea37f: (2) No such
file or directory
rbd: error open
I just yesterday updated by Ceph cluster from Quincy (17.2.7) to Reef (18.2.4,
purely by accident; I didn't realize I was getting it THE EXACT SECOND it was
pushed out).
The upgrade resolved an issue I was having with HTTP 500 errors on the RGW UI,
but seems to have created an issue in the dash
Sorry, the list has been wonky for me. I was logged in with my GitHub
credentials and when I try and publish the post, I get the message
This list is moderated, please subscribe to it before posting.
When I try and manage my subscription, I get:
Something went wrong
Mailman REST API not avai
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
I recently updated one of the hosts (an older Dell PowerEdge R515) in my Ceph
Quincy (17.2.6) cluster. I needed to change the IP address, so I removed the
host from the cluster (gracefully removed OSDs and daemons, then removed the
host). I also took the opportunity to upgrade the host from Rock
11 matches
Mail list logo