And sorry for all these mails, I forgot to mention that we are running
18.2.2.
Michel
Le 26/03/2025 à 21:51, Michel Jouvin a écrit :
Hi again,
Looking for more info on the degraded filesystem, I managed to connect
to the dashboard where I see an error not reported as explicitely by
'ceph health' :
One or more metadata daemons (MDS ranks) are failed or in a damaged
state. At best the filesystem is partially available, at worst the
filesystem is completely unusable.
But I don't manage what can be done from this point... and I really
don't understand how we ended up in such a state...
Cheers,
Michel
Le 26/03/2025 à 21:27, Michel Jouvin a écrit :
Hi,
We have a production cluster made of 3 mon+mgr, 18 OSD servers and
~500 OSDs and configured with ~50 pools, 1/2 EC (9+6) and 1/2 replica
3. It also has 2 CephFS filesystems with 1 MDS each.
2 days ago, in a period spanning 16 hours, 13 OSD crashed with an
OOM. The OSD were first restarted but it was decided to reboot the
server with a crashed OSD and "by mistake" (it was at least useless),
the OSD of the rebooted server were set noout,norebalance before the
reboot. The flags were removed after the reboot.
After all of this, 'ceph -s' started to report a lot of misplaced PG
and recovery started. All the PGs but one were successfully
reactivated. One stayed in the activating+remapped state (located in
a pool used for tests). 'ceph health' (I don't put the details here
to avoid a too long mail but I can shared them) says:
HEALTH_WARN 1 failed cephadm daemon(s); 1 filesystem is degraded; 2
MDSs report slow metadata IOs; Reduced data availability: 1 pg
inactive; 13 daemons have recently crashed
and reports about one of the filesystem being degraded despite the
only PG reported inactive is not part of a pool related to the FS.
The recovery was slow until we realized we should change the mclock
profile to high_recovery_ops. Then it completed in a few hours.
Unfortunately the degraded filesystem remains degraded without an
obvious reason... and the inactive page is still in the
activating+remapped state. We have not been able to identify a
relevant error in the logs up to now (but we may have missed
something...).
So far we have avoided restarting too many things until we have a
better understanding of what happened and what is the current state.
We only restarted the mgr which was using a lot of CPU and the MDS
for the degraded FS, without any improvement.
We are looking on advices about where to start... It seems we have
(at least) 2 independent problems:
- A PG that cannot be reactivated with a remap operation that doesn't
proceed: would stopping osd.17 help (so that osd.460 is reused)?
[root@ijc-mon1 ~]# ceph pg dump_stuck
PG_STAT STATE UP UP_PRIMARY ACTING
ACTING_PRIMARY
32.7ef activating+remapped [100,154,17] 100
[100,154,460] 100
- 1 degraded filesystem: where to look for a reason?
Thanks in advance for any help?
Cheers,
Michel
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io