Maybe a long shot, but have you checked OSD memory usage? Are the OSD hosts low on RAM and swapping to disk?
I am not familiar with your issue, but though that might cause it. Chris On 2016-11-14 3:29 pm, Brad Hubbard wrote: > Have you looked for clues in the output of dump_historic_ops ? > > On Tue, Nov 15, 2016 at 1:45 AM, Thomas Danan <thomas.da...@mycom-osi.com> > wrote: > > Thanks Luis, > > Here are some answers .... > > Journals are not on SSD and collocated with OSD daemons host. > > We look at the disk performances and did not notice anything wrong with > acceptable rw latency < 20ms. > > No issue on the network as well from what we have seen. > > There is only one pool in the cluster so pool size = cluster size. > Replication factor is default: 3 and there is no erasure coding. > > We tried to stop deep scrub but without notable effect. > > We have one near full OSD and adding new DN but I doubt this could be the > issue. > > I doubt we are hitting cluster limits but if it was the case, then adding new > DN should help. Also writing to primary OSD is working fine whereas writing > on secondary OSD is often blocked. Last is recovery can be very fast (several > GB/s) and seems never blocked, where client RW IOs are about several hundred > MB /s and are too much often blocked when writing replicas. > > Thomas > > FROM: Luis Periquito [mailto:periqu...@gmail.com] > SENT: lundi 14 novembre 2016 16:23 > TO: Thomas Danan > CC: ceph-users@lists.ceph.com > SUBJECT: Re: [ceph-users] ceph cluster having blocke requests very frequently > > Without knowing the cluster architecture it's hard to know exactly what may > be happening. > > How is the cluster hardware? Where are the journals? How busy are the disks > (% time busy)? What is the pool size? Are these replicated or EC pools? > > Have you tried tuning the deep-scrub processes? Have you tried stopping them > altogether? Are the journals on SSDs? As a first feeling the cluster may be > hitting it's limits (also you have at least one OSD getting full)... > > On Mon, Nov 14, 2016 at 3:16 PM, Thomas Danan <thomas.da...@mycom-osi.com> > wrote: > > Hi All, > > We have a cluster in production who is suffering from intermittent blocked > request (25 requests are blocked > 32 sec). The blocked request occurrences > are frequent and global to all OSDs. > > From the OSD daemon logs, I can see related messages: > > 16-11-11 18:25:29.917518 7fd28b989700 0 log_channel(cluster) log [WRN] : slow > request 30.429723 seconds old, received at 2016-11-11 18:24:59.487570: > osd_op(client.2406272.1:336025615 rbd_data.66e952ae8944a.0000000000350167 > [set-alloc-hint object_size 4194304 write_size 4194304,write 0~524288] > 0.8d3c9da5 snapc 248=[248,216] ondisk+write e201514) currently waiting for > subops from 210,499,821 > > . So I guess the issue is related to replication process when writing new > data on the cluster. Again it is never the same secondary OSDs that are > displayed in OSD daemon logs. > > As a result we are experiencing very important IO Write latency on ceph > client side (can be up to 1 hour !!!). > > We have checked Network health as well as disk health but we wre not able to > find any issue. > > Wanted to know if this issue was already observed or if you have ideas to > investigate / WA the issue. > > Many thanks... > > Thomas > > The cluster is composed with 37DN and 851 OSDs and 5 MONs > > The Ceph clients are accessing the client with RBD > > Cluster is Hammer 0.94.5 version > > cluster 1a26e029-3734-4b0e-b86e-ca2778d0c990 > > health HEALTH_WARN > > 25 requests are blocked > 32 sec > > 1 near full osd(s) > > noout flag(s) set > > monmap e3: 5 mons at > {NVMBD1CGK190D00=10.137.81.13:6789/0,nvmbd1cgy050d00=10.137.78.226:6789/0,nvmbd1cgy070d00=10.137.78.232:6789/0,nvmbd1cgy090d00=10.137.78.228:6789/0,nvmbd1cgy130d00=10.137.78.218:6789/0 > [1]} > > election epoch 664, quorum 0,1,2,3,4 > nvmbd1cgy130d00,nvmbd1cgy050d00,nvmbd1cgy090d00,nvmbd1cgy070d00,NVMBD1CGK190D00 > > > osdmap e205632: 851 osds: 850 up, 850 in > > flags noout > > pgmap v25919096: 10240 pgs, 1 pools, 197 TB data, 50664 kobjects > > 597 TB used, 233 TB / 831 TB avail > > 10208 active+clean > > 32 active+clean+scrubbing+deep > > client io 97822 kB/s rd, 205 MB/s wr, 2402 op/s > > THANK YOU > > THOMAS DANAN > > DIRECTOR OF PRODUCT DEVELOPMENT > > Office +33 1 49 03 77 53 [2] > > Mobile +33 7 76 35 76 43 [3] > > Skype thomas.danan > > www.mycom-osi.com [4] > > [4] > > Follow us on Twitter, LinkedIn, YouTube and our Blog > > [5] [6] [7] [8] > > ------------------------- > > This electronic message contains information from Mycom which may be > privileged or confidential. The information is intended to be for the use of > the individual(s) or entity named above. If you are not the intended > recipient, be aware that any disclosure, copying, distribution or any other > use of the contents of this information is prohibited. If you have received > this electronic message in error, please notify us by post or telephone (to > the numbers or correspondence address above) or by email (at the email > address above) immediately. > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [9] > > ------------------------- > > This electronic message contains information from Mycom which may be > privileged or confidential. The information is intended to be for the use of > the individual(s) or entity named above. If you are not the intended > recipient, be aware that any disclosure, copying, distribution or any other > use of the contents of this information is prohibited. If you have received > this electronic message in error, please notify us by post or telephone (to > the numbers or correspondence address above) or by email (at the email > address above) immediately. > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [9] -- Cheers, Brad _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [9] Links: ------ [1] http://10.137.81.13:6789/0,nvmbd1cgy050d00=10.137.78.226:6789/0,nvmbd1cgy070d00=10.137.78.232:6789/0,nvmbd1cgy090d00=10.137.78.228:6789/0,nvmbd1cgy130d00=10.137.78.218:6789/0 [2] tel:%2B33%201%2049%2003%2077%2053 [3] tel:%2B33%207%2076%2035%2076%2043 [4] http://www.mycom-osi.com/ [5] http://twitter.com/mycomosi [6] http://www.linkedin.com/company/mycom-osi [7] http://www.youtube.com/user/MYCOM-OSI [8] http://www.mycom-osi.com/blog [9] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com