You can't tell from the client log here, but probably the MDS itself was failing over to a new instance during that interval. There's not much experience with it, but you could experiment with faster failover by reducing the mds beacon and grace times. This may or may not work reliably...
On Sat, Feb 9, 2019 at 10:52 AM Fyodor Ustinov <u...@ufm.su> wrote: > Hi! > > I have ceph cluster with 3 nodes with mon/mgr/mds servers. > I reboot one node and see this in client log: > > Feb 09 20:29:14 ceph-nfs1 kernel: libceph: mon2 10.5.105.40:6789 socket > closed (con state OPEN) > Feb 09 20:29:14 ceph-nfs1 kernel: libceph: mon2 10.5.105.40:6789 session > lost, hunting for new mon > Feb 09 20:29:14 ceph-nfs1 kernel: libceph: mon0 10.5.105.34:6789 session > established > Feb 09 20:29:22 ceph-nfs1 kernel: libceph: mds0 10.5.105.40:6800 socket > closed (con state OPEN) > Feb 09 20:29:23 ceph-nfs1 kernel: libceph: mds0 10.5.105.40:6800 socket > closed (con state CONNECTING) > Feb 09 20:29:24 ceph-nfs1 kernel: libceph: mds0 10.5.105.40:6800 socket > closed (con state CONNECTING) > Feb 09 20:29:24 ceph-nfs1 kernel: libceph: mds0 10.5.105.40:6800 socket > closed (con state CONNECTING) > Feb 09 20:29:53 ceph-nfs1 kernel: ceph: mds0 reconnect start > Feb 09 20:29:53 ceph-nfs1 kernel: ceph: mds0 reconnect success > Feb 09 20:30:05 ceph-nfs1 kernel: ceph: mds0 recovery completed > > As I understand it, the following has happened: > 1. Client detects - link with mon server broken and fast switches to > another mon (less that 1 seconds). > 2. Client detects - link with mds server broken, 3 times trying reconnect > (unsuccessful), waiting and reconnects to the same mds after 30 seconds > downtime. > > I have 2 questions: > 1. Why? > 2. How to reduce switching time to another mds? > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com