Hi Xiubo,
Thanks for your analysis.
Is there anything I can do to put CephFS back in healthy state? Or
should I wait for to patch to fix that bug?
Cheers,
Adrien
Le 25/11/2022 à 06:13, Xiubo Li a écrit :
Hi Adren,
Thank you for your logs.
From your logs I found one bug and I have raised on
Hi!
osd_heartbeat_interval indicates interval (6 seconds) between peer pings,
if peer does not reply within osd_heartbeat_grace(20 seconds), osd will
report peer osd failure to mon, and then mon to mark down failure osd.
So, the client request will be blocked within 20 seconds, 20 seconds is too
Hello,
I have a question about osd removal/replacement:
I just removed an osd where the disk was still running but had read
errors, leading to failed deep scrubs - as the intent is to replace this
as soon as we manage to get a spare I removed it with the '--replace' flag:
# ceph orch osd rm 2
On 25/11/2022 16:25, Adrien Georget wrote:
Hi Xiubo,
Thanks for your analysis.
Is there anything I can do to put CephFS back in healthy state? Or
should I wait for to patch to fix that bug?
Please try to trim the journals and umount all the clients first, and
then to see could you pull up t