[ceph-users] Slow recovery and inaccurate recovery figures since Quincy upgrade

2023-10-02 Thread Iain Stott
Hi, We run 3 production clusters in a multi-site setup. They were deployed with Ceph-Ansible but recently switched to cephadm while on the Pacific release. Shortly after migrating to cephadm they were upgraded to Quincy. Since moving to Quincy, the recovery on one of the replica sites has tanke

[ceph-users] Re: cephfs health warn

2023-10-02 Thread Venky Shankar
Hi Ben, Are you using multimds without subtree pinning? On Tue, Oct 3, 2023 at 10:00 AM Ben wrote: > > Dear cephers: > more log captures(see below) show the full segments list(more than 3 to > be trimmed stuck, growing over time). any ideas to get out of this? > > Thanks, > Ben > > > debug 2

[ceph-users] Re: cephfs health warn

2023-10-02 Thread Ben
Dear cephers: more log captures(see below) show the full segments list(more than 3 to be trimmed stuck, growing over time). any ideas to get out of this? Thanks, Ben debug 2023-09-30T14:34:14.557+ 7f9c29bb1700 5 mds.4.log trim already expiring segment 195341004/893374309813, 180 events d

[ceph-users] Re: Clients failing to respond to capability release

2023-10-02 Thread E Taka
Same problem here with Ceph 17.2.6 on Ubuntu 22.04 and Clients Debian 11, Kernel 6.0.12-1~bpo11+1. We are still looking for a solution. At the time being we let restart the Orchestrator MDS daemons by removig/adding labels to the servers. We use multiple MDS and have many CPU cores and memory. The