On 20 December 2015 at 15:06, Francois Lafont <flafdiv...@free.fr> wrote:
> Hi, > > On 20/12/2015 19:47, Don Waterloo wrote: > > > I did a bit more work on this. > > > > On cephfs-fuse, I get ~700 iops. > > On cephfs kernel, I get ~120 iops. > > These were both on 4.3 kernel > > > > So i backed up to 3.16 kernel on the client. And observed the same > results. > > > > So ~20K iops w/ rbd, ~120iops w/ cephfs. > > Ok. Please, can you give us your configuration? > How many nodes, osds, ceph version, disks (SSD or not, HBA/controller), > RAM, CPU, network (1Gb/10Gb) etc.? All nodes have 10Gbps to each other OSD: $ ceph osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -1 5.48996 root default -2 0.89999 host nubo-1 0 0.89999 osd.0 up 1.00000 1.00000 -3 0.89999 host nubo-2 1 0.89999 osd.1 up 1.00000 1.00000 -4 0.89999 host nubo-3 2 0.89999 osd.2 up 1.00000 1.00000 -5 0.92999 host nubo-19 3 0.92999 osd.3 up 1.00000 1.00000 -6 0.92999 host nubo-20 4 0.92999 osd.4 up 1.00000 1.00000 -7 0.92999 host nubo-21 5 0.92999 osd.5 up 1.00000 1.00000 Each contains 1 x Samsung 850 Pro 1TB SSD (on sata) Each are Ubuntu 15.10 running 4.3.0-040300-generic kernel. Each are running ceph 0.94.5-0ubuntu0.15.10.1 nubo-1/nubo-2/nubo-3 are 2x X5650 @ 2.67GHz w/ 96GB ram. nubo-19/nubo-20/nubo-21 are 2x E5-2699 v3 @ 2.30GHz, w/ 576GB ram. the connections are to the chipset sata in each case. The fio test to the underlying xfs disk (e.g. cd /var/lib/ceph/osd/ceph-1; fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1 --name=readwrite --filename=rw.data --bs=4k --iodepth=64 --size=5000MB --readwrite=randrw --rwmixread=50) shows ~22K IOPS on each disk. nubo-1/2/3 are also the mon and the mds: $ ceph status cluster b23abffc-71c4-4464-9449-3f2c9fbe1ded health HEALTH_OK monmap e1: 3 mons at {nubo-1= 10.100.10.60:6789/0,nubo-2=10.100.10.61:6789/0,nubo-3=10.100.10.62:6789/0} election epoch 1104, quorum 0,1,2 nubo-1,nubo-2,nubo-3 mdsmap e621: 1/1/1 up {0=nubo-3=up:active}, 2 up:standby osdmap e2459: 6 osds: 6 up, 6 in pgmap v127331: 840 pgs, 6 pools, 144 GB data, 107 kobjects 289 GB used, 5332 GB / 5622 GB avail 840 active+clean client io 0 B/s rd, 183 kB/s wr, 54 op/s
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com