ok, I'll test it tomorrow , thank you.



------------------ Original ------------------
From:  "Irek Fasikhov";<malm...@gmail.com>;
Date:  Thu, Feb 12, 2015 09:29 PM
To:  "Alexandre DERUMIER"<aderum...@odiso.com>; 
Cc:  "killingwolf"<killingw...@qq.com>; 
"ceph-users"<ceph-users@lists.ceph.com>; 
Subject:  Re: [ceph-users] re: Upgrade 0.80.5 to 0.80.8 --the VM's read 
requestbecome too slow



Hi.
hmm ... I thought, why I have such a low speed reading on another cluster ....


P.S. ceph 0.80.8


2015-02-12 14:33 GMT+03:00 Alexandre DERUMIER <aderum...@odiso.com>:
>>Hi,
 >>Can you test with disabling rbd_cache ?
 
 >>I remember of a bug detected in giant, not sure it's also the case for 
 >>fireflt
 
 This was this tracker:
 
 http://tracker.ceph.com/issues/9513
 
 But It has been solved and backported to firefly.
 
 Also, can you test 0.80.6 and 0.80.7 ?
 
 
 
 
 
 
 
 ----- Mail original -----
 De: "killingwolf" <killingw...@qq.com>
 À: "ceph-users" <ceph-users@lists.ceph.com>
 Envoyé: Jeudi 12 Février 2015 12:16:32
 Objet: [ceph-users] re: Upgrade 0.80.5 to 0.80.8 --the VM's read requestbecome 
too slow
 
 I have this problems too , Help!
 
 ------------------ 原始邮件 ------------------
 发件人: "杨万元";<yangwanyuan8...@gmail.com>;
 发送时间: 2015年2月12日(星期四) 中午11:14
 收件人: "ceph-users@lists.ceph.com"<ceph-users@lists.ceph.com>;
 主题: [ceph-users] Upgrade 0.80.5 to 0.80.8 --the VM's read requestbecome too 
slow
 
 Hello!
 We use Ceph+Openstack in our private cloud. Recently we upgrade our centos6.5 
based cluster from Ceph Emperor to Ceph Firefly.
 At first,we use redhat yum repo epel to upgrade, this Ceph's version is 
0.80.5. First upgrade monitor,then osd,last client. when we complete this 
upgrade, we boot a VM on the cluster,then use fio to test the io performance. 
The io performance is as better as before. Everything is ok!
 Then we upgrade the cluster from 0.80.5 to 0.80.8,when we completed , we 
reboot the VM to load the newest librbd. after that we also use fio to test the 
io performance .then we find the randwrite and write is as good as before.but 
the randread and read is become worse, randwrite's iops from 4000-5000 to 
300-400 ,and the latency is worse. the write's bw from 400MB/s to 115MB/s . 
then I downgrade the ceph client version from 0.80.8 to 0.80.5, then the reslut 
become normal.
 So I think maybe something cause about librbd. I compare the 0.80.8 release 
notes with 0.80.5 ( http://ceph.com/docs/master/release-notes/#v0-80-8-firefly 
), I just find this change in 0.80.8 is something about read request : librbd: 
cap memory utilization for read requests (Jason Dillaman) . Who can explain 
this?
 
 
 My ceph cluster is 400osd,5mons :
 ceph -s
 health HEALTH_OK
 monmap e11: 5 mons at {BJ-M1-Cloud71= 
172.28.2.71:6789/0,BJ-M1-Cloud73=172.28.2.73:6789/0,BJ-M2-Cloud80=172.28.2.80:6789/0,BJ-M2-Cloud81=172.28.2.81:6789/0,BJ-M3-Cloud85=172.28.2.85:6789/0
 }, election epoch 198, quorum 0,1,2,3,4 
BJ-M1-Cloud71,BJ-M1-Cloud73,BJ-M2-Cloud80,BJ-M2-Cloud81,BJ-M3-Cloud85
 osdmap e120157: 400 osds: 400 up, 400 in
 pgmap v26161895: 29288 pgs, 6 pools, 20862 GB data, 3014 kobjects
 41084 GB used, 323 TB / 363 TB avail
 29288 active+clean
 client io 52640 kB/s rd, 32419 kB/s wr, 5193 op/s
 
 
 The follwing is my ceph client conf :
 [global]
 auth_service_required = cephx
 filestore_xattr_use_omap = true
 auth_client_required = cephx
 auth_cluster_required = cephx
 mon_host = 
172.29.204.24,172.29.204.48,172.29.204.55,172.29.204.58,172.29.204.73
 mon_initial_members = ZR-F5-Cloud24, ZR-F6-Cloud48, ZR-F7-Cloud55, 
ZR-F8-Cloud58, ZR-F9-Cloud73
 fsid = c01c8e28-304e-47a4-b876-cb93acc2e980
 mon osd full ratio = .85
 mon osd nearfull ratio = .75
 public network = 172.29.204.0/24
 mon warn on legacy crush tunables = false
 
 [osd]
 osd op threads = 12
 filestore journal writeahead = true
 filestore merge threshold = 40
 filestore split multiple = 8
 
 [client]
 rbd cache = true
 rbd cache writethrough until flush = false
 rbd cache size = 67108864
 rbd cache max dirty = 50331648
 rbd cache target dirty = 33554432
 
 [client.cinder]
 admin socket = /var/run/ceph/rbd-$pid.asok
 
 
 
 My VM is 8core16G,we use fio scripts is :
 fio -ioengine=libaio -bs=4k -direct=1 -thread -rw=randread -size=60G 
-filename=/dev/vdb -name="EBS" -iodepth=32 -runtime=200
 fio -ioengine=libaio -bs=4k -direct=1 -thread -rw=randwrite -size=60G 
-filename=/dev/vdb -name="EBS" -iodepth=32 -runtime=200
 fio -ioengine=libaio -bs=4k -direct=1 -thread -rw=read -size=60G 
-filename=/dev/vdb -name="EBS" -iodepth=32 -runtime=200
 fio -ioengine=libaio -bs=4k -direct=1 -thread -rw=write -size=60G 
-filename=/dev/vdb -name="EBS" -iodepth=32 -runtime=200
 
 The following is the io test result
 ceph client verison :0.80.5
 read: bw= 430MB
 write: bw=420MB
 

randread: iops= 4875 latency=65ms
 randwrite: iops=6844 latency=46ms
 
 ceph client verison :0.80.8
 read: bw= 115MB
 write: bw=480MB
 randread: iops= 381 latency=83ms
 randwrite: iops=4843 latency=68ms
 
 _______________________________________________
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
 _______________________________________________
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
 




-- 
С уважением, Фасихов Ирек НургаязовичМоб.: +79229045757
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to