On Fri, Jun 13, 2014 at 5:50 PM, Ke-fei Lin <k...@kfei.net> wrote: > 2014-06-13 21:23 GMT+08:00 Andrey Korolyov <and...@xdel.ru>: > >> On Fri, Jun 13, 2014 at 7:09 AM, Ke-fei Lin <k...@kfei.net> wrote: >> > Hi list, >> > >> > I deployed a Windows 7 VM with qemu-rbd disk, and got an unexpected >> > booting >> > phase performance. >> > >> > I discovered that when booting the Windows VM up, there are consecutive >> > ~2 >> > minutes that `ceph -w` gives me an interesting log like: "... 567 KB/s >> > rd, >> > 567 op/s", "... 789 KB/s rd, 789 op/s" and so on. >> > >> > e.g. >> > 2014-06-05 15:47:43.125441 mon.0 [INF] pgmap v18095: 320 pgs: 320 >> > active+clean; 86954 MB data, 190 GB used, 2603 GB / 2793 GB avail; 765 >> > kB/s >> > rd, 765 op/s >> > 2014-06-05 15:47:44.240662 mon.0 [INF] pgmap v18096: 320 pgs: 320 >> > active+clean; 86954 MB data, 190 GB used, 2603 GB / 2793 GB avail; 568 >> > kB/s >> > rd, 568 op/s >> > ... (skipped) >> > 2014-06-05 15:50:02.441523 mon.0 [INF] pgmap v18186: 320 pgs: 320 >> > active+clean; 86954 MB data, 190 GB used, 2603 GB / 2793 GB avail; 412 >> > kB/s >> > rd, 412 op/s >> > >> > Which shows the number of rps is always the same as the number of ops, >> > i.e. >> > every operation is nearly 1KB, and I think this leads a very long boot >> > time >> > (takes 2 mins to enter desktop). But I can't understand why, is it an >> > issue >> > of my Ceph cluster? Or just some special I/O patterns in Windows VM >> > booting >> > process? >> > >> > In addition, I know that there are no qemu-rbd caching benefits during >> > boot >> > phase since the cache is not persistent (please corrects me), so is it >> > possible to enlarge the read_ahead size in qemu-rbd driver? And does >> > this >> > make any sense? >> > >> > And finally, how can I tune up my Ceph cluster for this workload >> > (booting >> > Windows VM)? >> > >> > Any advice and suggestions will be greatly appreciated. >> > >> > >> > Context: >> > >> > 4 OSDs (7200rpm/750GB/SATA) with replication factor 2. >> > >> > The system disk in Windows VM is NTFS formatted with default 4K block >> > size. >> > >> > $ uname -a >> > Linux ceph-consumer 3.11.0-22-generic #38~precise1-Ubuntu SMP Fri >> > May 16 >> > 20:47:57 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux >> > >> > $ ceph --version >> > ceph version 0.80.1 (a38fe1169b6d2ac98b427334c12d7cf81f809b74) >> > >> > $ dpkg -l | grep rbd >> > ii librbd-dev 0.80.1-1precise >> > RADOS block device client library (development files) >> > ii librbd1 0.80.1-1precise >> > RADOS block device client library >> > >> > $ virsh version >> > Compiled against library: libvir 0.9.8 >> > Using library: libvir 0.9.8 >> > Using API: QEMU 0.9.8 >> > Running hypervisor: QEMU 1.7.1 () >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > >> >> Hi, >> >> If you are able to leave only this VM in cluster scope to check, >> you`ll perhaps may use virsh domblkstat accumulated values to compare >> real number of operations. > > > Thanks, Andrey. > > I tried `virsh domblkstat <vm> hda` (only this VM in whole cluster) and got > these values: > > hda rd_req 70682 > hda rd_bytes 229894656 > hda wr_req 1067 > hda wr_bytes 12645888 > hda flush_operations 0 > > (These values became stable after ~2 mins) > > While the output of `ceph -w` is attached at: http://pastebin.com/Uhdj9drV > > Any advices?
Thanks, poor man`s analysis shows that it can be true - assuming median heartbeat value as 1.2s, overall read ops are about 40k, which is close enough to what qemu stats saying, regarding floating heartbeat interval. Because ceph -w never had such value as a precise measurement tool, I may suggest to measure block stats difference on smaller intervals, about 1s or so, and compare values then. By the way, what driver do you use in qemu for a block device? _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com