Hi,
We run 3 production clusters in a multi-site setup. They were deployed with
Ceph-Ansible but recently switched to cephadm while on the Pacific release.
Shortly after migrating to cephadm they were upgraded to Quincy. Since moving
to Quincy, the recovery on one of the replica sites has tanke
Hi Ben,
Are you using multimds without subtree pinning?
On Tue, Oct 3, 2023 at 10:00 AM Ben wrote:
>
> Dear cephers:
> more log captures(see below) show the full segments list(more than 3 to
> be trimmed stuck, growing over time). any ideas to get out of this?
>
> Thanks,
> Ben
>
>
> debug 2
Dear cephers:
more log captures(see below) show the full segments list(more than 3 to
be trimmed stuck, growing over time). any ideas to get out of this?
Thanks,
Ben
debug 2023-09-30T14:34:14.557+ 7f9c29bb1700 5 mds.4.log trim already
expiring segment 195341004/893374309813, 180 events
d
Same problem here with Ceph 17.2.6 on Ubuntu 22.04 and Clients Debian 11,
Kernel 6.0.12-1~bpo11+1.
We are still looking for a solution. At the time being we let restart the
Orchestrator MDS daemons by removig/adding labels to the servers. We use
multiple MDS and have many CPU cores and memory. The