Maybe a long shot, but have you checked OSD memory usage? Are the OSD
hosts low on RAM and swapping to disk? 

I am not familiar with your issue, but though that might cause it. 

Chris 

On 2016-11-14 3:29 pm, Brad Hubbard wrote: 

> Have you looked for clues in the output of dump_historic_ops ? 
> 
> On Tue, Nov 15, 2016 at 1:45 AM, Thomas Danan <thomas.da...@mycom-osi.com> 
> wrote:
> 
> Thanks Luis, 
> 
> Here are some answers .... 
> 
> Journals are not on SSD and collocated with OSD daemons host. 
> 
> We look at the disk performances and did not notice anything wrong with 
> acceptable rw latency < 20ms. 
> 
> No issue on the network as well from what we have seen. 
> 
> There is only one pool in the cluster so pool size = cluster size. 
> Replication factor is default: 3 and there is no erasure coding. 
> 
> We tried to stop deep scrub but without notable effect. 
> 
> We have one near full OSD and adding new DN but I doubt this could be the 
> issue. 
> 
> I doubt we are hitting cluster limits but if it was the case, then adding new 
> DN should help. Also writing to primary OSD is working fine whereas writing 
> on secondary OSD is often blocked. Last is recovery can be very fast (several 
> GB/s) and seems never blocked, where client RW IOs are about several hundred 
> MB /s and are too much often blocked when writing replicas. 
> 
> Thomas 
> 
> FROM: Luis Periquito [mailto:periqu...@gmail.com] 
> SENT: lundi 14 novembre 2016 16:23
> TO: Thomas Danan
> CC: ceph-users@lists.ceph.com
> SUBJECT: Re: [ceph-users] ceph cluster having blocke requests very frequently 
> 
> Without knowing the cluster architecture it's hard to know exactly what may 
> be happening. 
> 
> How is the cluster hardware? Where are the journals? How busy are the disks 
> (% time busy)? What is the pool size? Are these replicated or EC pools? 
> 
> Have you tried tuning the deep-scrub processes? Have you tried stopping them 
> altogether? Are the journals on SSDs? As a first feeling the cluster may be 
> hitting it's limits (also you have at least one OSD getting full)... 
> 
> On Mon, Nov 14, 2016 at 3:16 PM, Thomas Danan <thomas.da...@mycom-osi.com> 
> wrote: 
> 
> Hi All, 
> 
> We have a cluster in production who is suffering from intermittent blocked 
> request (25 requests are blocked > 32 sec). The blocked request occurrences 
> are frequent and global to all OSDs. 
> 
> From the OSD daemon logs, I can see related messages: 
> 
> 16-11-11 18:25:29.917518 7fd28b989700 0 log_channel(cluster) log [WRN] : slow 
> request 30.429723 seconds old, received at 2016-11-11 18:24:59.487570: 
> osd_op(client.2406272.1:336025615 rbd_data.66e952ae8944a.0000000000350167 
> [set-alloc-hint object_size 4194304 write_size 4194304,write 0~524288] 
> 0.8d3c9da5 snapc 248=[248,216] ondisk+write e201514) currently waiting for 
> subops from 210,499,821 
> 
> . So I guess the issue is related to replication process when writing new 
> data on the cluster. Again it is never the same secondary OSDs that are 
> displayed in OSD daemon logs. 
> 
> As a result we are experiencing very important IO Write latency on ceph 
> client side (can be up to 1 hour !!!). 
> 
> We have checked Network health as well as disk health but we wre not able to 
> find any issue. 
> 
> Wanted to know if this issue was already observed or if you have ideas to 
> investigate / WA the issue. 
> 
> Many thanks... 
> 
> Thomas 
> 
> The cluster is composed with 37DN and 851 OSDs and 5 MONs 
> 
> The Ceph clients are accessing the client with RBD 
> 
> Cluster is Hammer 0.94.5 version 
> 
> cluster 1a26e029-3734-4b0e-b86e-ca2778d0c990 
> 
> health HEALTH_WARN 
> 
> 25 requests are blocked > 32 sec 
> 
> 1 near full osd(s) 
> 
> noout flag(s) set 
> 
> monmap e3: 5 mons at 
> {NVMBD1CGK190D00=10.137.81.13:6789/0,nvmbd1cgy050d00=10.137.78.226:6789/0,nvmbd1cgy070d00=10.137.78.232:6789/0,nvmbd1cgy090d00=10.137.78.228:6789/0,nvmbd1cgy130d00=10.137.78.218:6789/0
>  [1]} 
> 
> election epoch 664, quorum 0,1,2,3,4 
> nvmbd1cgy130d00,nvmbd1cgy050d00,nvmbd1cgy090d00,nvmbd1cgy070d00,NVMBD1CGK190D00
>  
> 
> osdmap e205632: 851 osds: 850 up, 850 in 
> 
> flags noout 
> 
> pgmap v25919096: 10240 pgs, 1 pools, 197 TB data, 50664 kobjects 
> 
> 597 TB used, 233 TB / 831 TB avail 
> 
> 10208 active+clean 
> 
> 32 active+clean+scrubbing+deep 
> 
> client io 97822 kB/s rd, 205 MB/s wr, 2402 op/s 
> 
> THANK YOU 
> 
> THOMAS DANAN 
> 
> DIRECTOR OF PRODUCT DEVELOPMENT 
> 
> Office +33 1 49 03 77 53 [2] 
> 
> Mobile +33 7 76 35 76 43 [3] 
> 
> Skype thomas.danan 
> 
> www.mycom-osi.com [4] 
> 
> [4] 
> 
> Follow us on Twitter, LinkedIn, YouTube and our Blog 
> 
> [5] [6] [7] [8] 
> 
> -------------------------
> 
> This electronic message contains information from Mycom which may be 
> privileged or confidential. The information is intended to be for the use of 
> the individual(s) or entity named above. If you are not the intended 
> recipient, be aware that any disclosure, copying, distribution or any other 
> use of the contents of this information is prohibited. If you have received 
> this electronic message in error, please notify us by post or telephone (to 
> the numbers or correspondence address above) or by email (at the email 
> address above) immediately. 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [9] 
> 
> -------------------------
> 
> This electronic message contains information from Mycom which may be 
> privileged or confidential. The information is intended to be for the use of 
> the individual(s) or entity named above. If you are not the intended 
> recipient, be aware that any disclosure, copying, distribution or any other 
> use of the contents of this information is prohibited. If you have received 
> this electronic message in error, please notify us by post or telephone (to 
> the numbers or correspondence address above) or by email (at the email 
> address above) immediately.
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [9]

-- 

Cheers,
Brad 

_______________________________________________
 ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [9] 

Links:
------
[1]
http://10.137.81.13:6789/0,nvmbd1cgy050d00=10.137.78.226:6789/0,nvmbd1cgy070d00=10.137.78.232:6789/0,nvmbd1cgy090d00=10.137.78.228:6789/0,nvmbd1cgy130d00=10.137.78.218:6789/0
[2] tel:%2B33%201%2049%2003%2077%2053
[3] tel:%2B33%207%2076%2035%2076%2043
[4] http://www.mycom-osi.com/
[5] http://twitter.com/mycomosi
[6] http://www.linkedin.com/company/mycom-osi
[7] http://www.youtube.com/user/MYCOM-OSI
[8] http://www.mycom-osi.com/blog
[9] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to