Hi,
I have not seen an answer yet, help would be very much appreciated as
our production cluster seems in a worst shape that initially described...
After a deeper analysis, we found that more than half of the pools,
despite reported as ok, are not accessible: the 'rados ls' command is
stuck when we try to access them. It is not correlated to the EC versus
3 replica pool configuration (both are affected or can be ok). We don't
have clear view whether we have a corruption problem (but it is unclear
why it would have happened) or a communication problem between some
cluster components that may explain that pools reported as good are not
accessible and that one PG remains in the activating+remapped state,
something that seems unusual (I could not find any reference to this
with my Google searches).
Then we are hesitating between forcing a deep scrub of everything (we
have not seen any errors reported by scrubs/deep scrubs run in the last
days) or restarting the whole cluster in case there is a kind of
deadlock in the communication between some mon and/or osd. In case the
second approach is suggested (cluster restart), should we do it daemon
by daemon or shutdown everything and do a cold restart of the cluster?
I don't want to clutter this already too long thread with too many
details but one of my colleague gave me the 'ceph -s' and 'ceph osd
status` output before he started to reboot servers. If it is useful, I
can share them.
Again, thanks in advance for any help/hint.
Best regards,
Michel
Le 26/03/2025 à 21:54, Michel Jouvin a écrit :
And sorry for all these mails, I forgot to mention that we are running
18.2.2.
Michel
Le 26/03/2025 à 21:51, Michel Jouvin a écrit :
Hi again,
Looking for more info on the degraded filesystem, I managed to
connect to the dashboard where I see an error not reported as
explicitely by 'ceph health' :
One or more metadata daemons (MDS ranks) are failed or in a damaged
state. At best the filesystem is partially available, at worst the
filesystem is completely unusable.
But I don't manage what can be done from this point... and I really
don't understand how we ended up in such a state...
Cheers,
Michel
Le 26/03/2025 à 21:27, Michel Jouvin a écrit :
Hi,
We have a production cluster made of 3 mon+mgr, 18 OSD servers and
~500 OSDs and configured with ~50 pools, 1/2 EC (9+6) and 1/2
replica 3. It also has 2 CephFS filesystems with 1 MDS each.
2 days ago, in a period spanning 16 hours, 13 OSD crashed with an
OOM. The OSD were first restarted but it was decided to reboot the
server with a crashed OSD and "by mistake" (it was at least
useless), the OSD of the rebooted server were set noout,norebalance
before the reboot. The flags were removed after the reboot.
After all of this, 'ceph -s' started to report a lot of misplaced PG
and recovery started. All the PGs but one were successfully
reactivated. One stayed in the activating+remapped state (located in
a pool used for tests). 'ceph health' (I don't put the details here
to avoid a too long mail but I can shared them) says:
HEALTH_WARN 1 failed cephadm daemon(s); 1 filesystem is degraded; 2
MDSs report slow metadata IOs; Reduced data availability: 1 pg
inactive; 13 daemons have recently crashed
and reports about one of the filesystem being degraded despite the
only PG reported inactive is not part of a pool related to the FS.
The recovery was slow until we realized we should change the mclock
profile to high_recovery_ops. Then it completed in a few hours.
Unfortunately the degraded filesystem remains degraded without an
obvious reason... and the inactive page is still in the
activating+remapped state. We have not been able to identify a
relevant error in the logs up to now (but we may have missed
something...).
So far we have avoided restarting too many things until we have a
better understanding of what happened and what is the current state.
We only restarted the mgr which was using a lot of CPU and the MDS
for the degraded FS, without any improvement.
We are looking on advices about where to start... It seems we have
(at least) 2 independent problems:
- A PG that cannot be reactivated with a remap operation that
doesn't proceed: would stopping osd.17 help (so that osd.460 is
reused)?
[root@ijc-mon1 ~]# ceph pg dump_stuck
PG_STAT STATE UP UP_PRIMARY
ACTING ACTING_PRIMARY
32.7ef activating+remapped [100,154,17] 100
[100,154,460] 100
- 1 degraded filesystem: where to look for a reason?
Thanks in advance for any help?
Cheers,
Michel
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io