Hi Blairo, > fio shows 70 MB/s seq read with 4M blocks, libaio, 1 thread, direct. > > fio seq write 200 MB/s > > The fio numbers are from fio running on a CephFS mount I take it?
exactly. > > # rados bench -t 1 -p test 60 write --no-cleanup > > I don't see an rbd test anywhere here...? > > I suggest comparing fio on CephFS with fio on rbd (as in using fio's > rbd ioengine), then at least the application side of your tests is > constant. > I tested four access types: 1. rbd kernel module, xfs, fio-libaio 2. rados bench seq 3. fio-librbd 4. ceph kernel module, fio-libaio Done tests, no big diference, all methods give 50 to 80 MB/s single threaded bandwidth. In my setup I observe rados bench results very similar to kernel module rbd results while fio-librbd having 1.5 times lower bandwidth (maybe because it runs in userspace?) What can be tuned to improve sequential read? readahead settings are: "client_readahead_min": "131072", "client_readahead_max_bytes": "2097152", "client_readahead_max_periods": "4", "rbd_readahead_trigger_requests": "10", "rbd_readahead_max_bytes": "524288", "rbd_readahead_disable_after_bytes": "52428800", -- WBR, Ilja.
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com