On Fri, Jun 30, 2017 at 2:14 PM, Nick Fisk <n...@fisk.me.uk> wrote: > > >> -----Original Message----- >> From: Ilya Dryomov [mailto:idryo...@gmail.com] >> Sent: 29 June 2017 18:54 >> To: Nick Fisk <n...@fisk.me.uk> >> Cc: Ceph Users <ceph-users@lists.ceph.com> >> Subject: Re: [ceph-users] Kernel mounted RBD's hanging >> >> On Thu, Jun 29, 2017 at 6:22 PM, Nick Fisk <n...@fisk.me.uk> wrote: >> >> -----Original Message----- >> >> From: Ilya Dryomov [mailto:idryo...@gmail.com] >> >> Sent: 29 June 2017 16:58 >> >> To: Nick Fisk <n...@fisk.me.uk> >> >> Cc: Ceph Users <ceph-users@lists.ceph.com> >> >> Subject: Re: [ceph-users] Kernel mounted RBD's hanging >> >> >> >> On Thu, Jun 29, 2017 at 4:30 PM, Nick Fisk <n...@fisk.me.uk> wrote: >> >> > Hi All, >> >> > >> >> > Putting out a call for help to see if anyone can shed some light on >> >> > this. >> >> > >> >> > Configuration: >> >> > Ceph cluster presenting RBD's->XFS->NFS->ESXi Running 10.2.7 on the >> >> > OSD's and 4.11 kernel on the NFS gateways in a pacemaker cluster >> >> > Both OSD's and clients are go into a pair of switches, single L2 >> >> > domain (no sign from pacemaker that there is network connectivity >> >> > issues) >> >> > >> >> > Symptoms: >> >> > - All RBD's on a single client randomly hang for 30s to several >> >> > minutes, confirmed by pacemaker and ESXi hosts complaining >> >> >> >> Hi Nick, >> >> >> >> What is a "single client" here? >> > >> > I mean a node of the pacemaker cluster. So all RBD's on the same >> pacemaker node hang. >> > >> >> >> >> > - Cluster load is minimal when this happens most times >> >> >> >> Can you post gateway syslog and point at when this happened? >> >> Corresponding pacemaker excerpts won't hurt either. >> > >> > Jun 28 16:35:38 MS-CEPH-Proxy1 lrmd[2026]: warning: p_export_ceph- >> ds1_monitor_60000 process (PID 17754) timed out >> > Jun 28 16:35:43 MS-CEPH-Proxy1 lrmd[2026]: crit: p_export_ceph- >> ds1_monitor_60000 process (PID 17754) will not die! >> > Jun 28 16:43:51 MS-CEPH-Proxy1 lrmd[2026]: warning: >> > p_export_ceph-ds1_monitor_60000:17754 - timed out after 30000ms Jun >> 28 16:43:52 MS-CEPH-Proxy1 IPaddr(p_vip_ceph-ds1)[28482]: INFO: ifconfig >> ens224:0 down >> > Jun 28 16:43:52 MS-CEPH-Proxy1 lrmd[2026]: notice: p_vip_ceph- >> ds1_stop_0:28482:stderr [ SIOCDELRT: No such process ] >> > Jun 28 16:43:52 MS-CEPH-Proxy1 crmd[2029]: notice: Operation >> p_vip_ceph-ds1_stop_0: ok (node=MS-CEPH-Proxy1, call=471, rc=0, cib- >> update=318, confirmed=true) >> > Jun 28 16:43:52 MS-CEPH-Proxy1 exportfs(p_export_ceph-ds1)[28499]: >> INFO: Un-exporting file system ... >> > Jun 28 16:43:52 MS-CEPH-Proxy1 exportfs(p_export_ceph-ds1)[28499]: >> > INFO: unexporting 10.3.20.0/24:/mnt/Ceph-DS1 Jun 28 16:43:52 >> > MS-CEPH-Proxy1 exportfs(p_export_ceph-ds1)[28499]: INFO: Unlocked NFS >> export /mnt/Ceph-DS1 Jun 28 16:43:52 MS-CEPH-Proxy1 >> exportfs(p_export_ceph-ds1)[28499]: INFO: Un-exported file system(s) >> > Jun 28 16:43:52 MS-CEPH-Proxy1 crmd[2029]: notice: Operation >> p_export_ceph-ds1_stop_0: ok (node=MS-CEPH-Proxy1, call=473, rc=0, cib- >> update=319, confirmed=true) >> > Jun 28 16:43:52 MS-CEPH-Proxy1 exportfs(p_export_ceph-ds1)[28549]: >> INFO: Exporting file system(s) ... >> > Jun 28 16:43:52 MS-CEPH-Proxy1 exportfs(p_export_ceph-ds1)[28549]: >> > INFO: exporting 10.3.20.0/24:/mnt/Ceph-DS1 Jun 28 16:43:52 MS-CEPH- >> Proxy1 exportfs(p_export_ceph-ds1)[28549]: INFO: directory /mnt/Ceph-DS1 >> exported >> > Jun 28 16:43:52 MS-CEPH-Proxy1 crmd[2029]: notice: Operation >> p_export_ceph-ds1_start_0: ok (node=MS-CEPH-Proxy1, call=474, rc=0, cib- >> update=320, confirmed=true) >> > >> > If I enable the read/write checks for the FS resource, they also timeout at >> the same time. >> >> What about syslog that the above corresponds to? > > I get exactly the same "_monitor" timeout message.
No "libceph: " or "rbd: " messages at all? No WARNs or hung tasks? > > Is there anything logging wise I can do with the kernel client to log when an > IO is taking a long time. Sort of like the slow requests in Ceph, but client > side? Nothing out of the box, as slow requests are usually not the client implementation's fault. Can you put together a script that would snapshot all files in /sys/kernel/debug/ceph/<cluster-fsid.client-id>/* on the gateways every second and rotate on an hourly basis? One of those files, osdc, lists in-flight requests. If that's empty when the timeouts occur then it's probably not krbd. What Maged said, and also can you clarify what those "read/write checks for the FS resource" do exactly? read/write to local xfs on /dev/rbd* or further up? Thanks, Ilya _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com