If I use slow HDD, I can get the same outcome. Placing journals on fast SAS or 
NVMe SSD will make a difference. If you are using SATA SSD, those SSD are much 
slower. Instead of guessing why Ceph is lagging, have you looked at ceph -w and 
iostat and vmstat reports during your tests? Io stat will tell you HDD and SSD 
stats (i use the commands: iostat -tzxm 5 to show only active disks). If they 
are dedicated luns, then look at %utilization and service times. Looking at 
vmstat, check the ‘b’ column which shows how often your system is blocked 
waiting on io.

> On Nov 25, 2016, at 8:48 AM, Kevin Olbrich <k...@sv01.de> wrote:
> 
> Hi,
> 
> we are running 80 VMs using KVM in OpenStack via RBD in Ceph Jewel on a total 
> of 53 disks (RAID parity already excluded).
> Our nodes are using Intel P3700 DC-SSDs for journaling.
> 
> Most VMs are linux based and load is low to medium. There are also about 10 
> VMs running Windows 2012R2, two of them run remote services (terminal).
> 
> My question is: Are 80 VMs hosted on 53 disks (mostly 7.2k SATA) to much? We 
> sometime experience lags where nearly all servers suffer from "blocked IO > 
> 32" seconds.
> 
> What are your experiences?
> 
> Mit freundlichen Grüßen / best regards,
> Kevin Olbrich.
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Rick Stehno


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to