You have 2 drives at almost 100% util which means they are maxed. So you need more disks or better drives to fix your io issues (SSDs for MySQL is a no brainer really) ------ Ursprungligt meddelande------Från: Sam HuracanDatum: lör 24 mars 2018 19:20Till: c...@elchaka.de;Kopia: ceph-users@lists.ceph.com;Ämne:Re: [ceph-users] Fwd: High IOWait Issue This is from iostat: I'm using Ceph jewel, has no HW error.Ceph health OK, we've just use 50% total volume.
2018-03-24 22:20 GMT+07:00 <c...@elchaka.de>: I would Check with Tools like atop the utilization of your Disks also. Perhaps something Related in dmesg or dorthin? - Mehmet Am 24. März 2018 08:17:44 MEZ schrieb Sam Huracan <nowitzki.sa...@gmail.com>: Hi guys,We are running a production OpenStack backend by Ceph. At present, we are meeting an issue relating to high iowait in VM, in some MySQL VM, we see sometime IOwait reaches abnormal high peaks which lead to slow queries increase, despite load is stable (we test with script simulate real load), you can see in graph.https://prnt.sc/ivndni MySQL VM are place on Ceph HDD Cluster, with 1 SSD journal for 7 HDD. In this cluster, IOwait on each ceph host is about 20%.https://prnt.sc/ivne08 Can you guy help me find the root cause of this issue, and how to eliminate this high iowait? Thanks in advance. _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com