[ceph-users] Re: filesystem became read only after Quincy upgrade

2022-11-25 Thread Adrien Georget
Hi Xiubo, Thanks for your analysis. Is there anything I can do to put CephFS back in healthy state? Or should I wait for to patch to fix that bug? Cheers, Adrien Le 25/11/2022 à 06:13, Xiubo Li a écrit : Hi Adren, Thank you for your logs. From your logs I found one bug and I have raised on

[ceph-users] Is there any risk in adjusting the osd_heartbeat_grace & osd_heartbeat_interval

2022-11-25 Thread yite gu
Hi! osd_heartbeat_interval indicates interval (6 seconds) between peer pings, if peer does not reply within osd_heartbeat_grace(20 seconds), osd will report peer osd failure to mon, and then mon to mark down failure osd. So, the client request will be blocked within 20 seconds, 20 seconds is too

[ceph-users] osd removal leaves 'stray daemon'

2022-11-25 Thread Holger Naundorf
Hello, I have a question about osd removal/replacement: I just removed an osd where the disk was still running but had read errors, leading to failed deep scrubs - as the intent is to replace this as soon as we manage to get a spare I removed it with the '--replace' flag: # ceph orch osd rm 2

[ceph-users] Re: filesystem became read only after Quincy upgrade

2022-11-25 Thread Xiubo Li
On 25/11/2022 16:25, Adrien Georget wrote: Hi Xiubo, Thanks for your analysis. Is there anything I can do to put CephFS back in healthy state? Or should I wait for to patch to fix that bug? Please try to trim the journals and umount all the clients first, and then to see could you pull up t