Hi Gregory, The latency is under my ceph test environment. It has only 1 host with 16 2T SATA disks(16 OSDs) and 10Gb/s NIC, journal and data are on the same disk which fs type is ext4. my cluster config is like this. is it normal under this configuration ? or how can i improve performance ?
Thanks. [root@storage1 ~]# ceph -s cluster 3429fd17-4a92-4d3b-a7fa-04adedb0da82 health HEALTH_OK monmap e1: 1 mons at {storage1=193.168.1.100:6789/0}, election epoch 1, quorum 0 storage1 osdmap e373: 16 osds: 16 up, 16 in pgmap v3515: 1024 pgs, 1 pools, 15635 MB data, 3946 objects 87220 MB used, 27764 GB / 29340 GB avail 1024 active+clean [root@storage1 ~]# ceph osd tree # id weight type name up/down reweight -1 16 root default -2 16 host storage1 0 1 osd.0 up 1 1 1 osd.1 up 1 2 1 osd.2 up 1 3 1 osd.3 up 1 4 1 osd.4 up 1 5 1 osd.5 up 1 6 1 osd.6 up 1 7 1 osd.7 up 1 8 1 osd.8 up 1 9 1 osd.9 up 1 10 1 osd.10 up 1 11 1 osd.11 up 1 12 1 osd.12 up 1 13 1 osd.13 up 1 14 1 osd.14 up 1 15 1 osd.15 up 1 [root@storage1 ~]# cat /etc/ceph/ceph.conf [global] fsid = 3429fd17-4a92-4d3b-a7fa-04adedb0da82 public network = 193.168.1.0/24 auth cluster required = cephx auth service required = cephx auth client required = cephx filestore xattr use omap = true [mon] debug paxos = 0/5 [mon.storage1] host = storage1 mon addr = 193.168.1.100:6789 [osd.0] host = storage1 [osd.1] host = storage1 [osd.2] host = storage1 [osd.3] host = storage1 [osd.4] host = storage1 [osd.5] host = storage1 [osd.6] host = storage1 [osd.7] host = storage1 [osd.8] host = storage1 [osd.9] host = storage1 [osd.10] host = storage1 [osd.11] host = storage1 [osd.12] host = storage1 [osd.13] host = storage1 [osd.14] host = storage1 [osd.15] host = storage1 Re: [ceph-users] why my "fs_commit_latency" is so high ? is it normal ? Gregory Farnum 收件人: duan.xufeng 2014/03/17 13:08 抄送: "ceph-users@lists.ceph.com" That seems a little high; how do you have your system configured? That latency is how long it takes for the hard drive to durably write out something to the journal. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sun, Mar 16, 2014 at 9:59 PM, <duan.xuf...@zte.com.cn> wrote: > > [root@storage1 ~]# ceph osd perf > osdid fs_commit_latency(ms) fs_apply_latency(ms) > 0 149 52 > 1 201 61 > 2 176 166 > 3 240 57 > 4 219 49 > 5 167 56 > 6 201 54 > 7 175 188 > 8 192 124 > 9 367 193 > 10 343 160 > 11 183 110 > 12 158 143 > 13 267 147 > 14 150 155 > 15 159 54 > > -------------------------------------------------------- > ZTE Information Security Notice: The information contained in this mail (and > any attachment transmitted herewith) is privileged and confidential and is > intended for the exclusive use of the addressee(s). If you are not an > intended recipient, any disclosure, reproduction, distribution or other > dissemination or use of the information contained is strictly prohibited. > If you have received this mail in error, please delete it and notify us > immediately. > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -------------------------------------------------------- ZTE Information Security Notice: The information contained in this mail (and any attachment transmitted herewith) is privileged and confidential and is intended for the exclusive use of the addressee(s). If you are not an intended recipient, any disclosure, reproduction, distribution or other dissemination or use of the information contained is strictly prohibited. If you have received this mail in error, please delete it and notify us immediately.
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com