Hi, As mentioned in my previous emails, I'm extremely new to ceph, so please forgive my lack of knowledge.
I'm trying to find a good way to mount ceph rbd images for export by LIO/targetcli rbd-nbd isn't good as it stops at 16 block devices (/dev/nbd0-15) kernel rbd mapping doesn't have support for new features. I thought rbd-fuse looked good, except write performance is abysmal. rados bench gives me ~250MB/s of write speed. an image mounted with rbd-fuse gives me ~2MB/s of write speed. CephFS write speeds are good as well. Is something wrong with my testing method or configuration? root@stor-vm1:/# ceph osd pool create rbd_storage 128 128 root@stor-vm1:/# rbd create --pool=rbd_storage --size=25G rbd_25g1 root@stor-vm1:/# mkdir /mnt/rbd root@stor-vm1:/# cd /mnt root@stor-vm1:/# rbd-fuse rbd -p rbd_storage root@stor-vm1:/# cd rbd root@stor-vm1:/# dd if=/dev/zero of=rbd_25g1 bs=4M count=2 status=progress 8388608 bytes (8.4 MB, 8.0 MiB) copied, 4.37754 s, 1.9 MB/s 2+0 records in 2+0 records out 8388608 bytes (8.4 MB, 8.0 MiB) copied, 4.3776 s, 1.9 MB/s rados bench: root@stor-vm1:/mnt/rbd# rados bench -p rbd_storage 10 write 2017-06-27 18:56:59.505647 7fb9c24a7e00 -1 WARNING: the following dangerous and experimental features are enabled: bluestore 2017-06-27 18:56:59.505768 7fb9c24a7e00 -1 WARNING: the following dangerous and experimental features are enabled: bluestore 2017-06-27 18:56:59.507385 7fb9c24a7e00 -1 WARNING: the following dangerous and experimental features are enabled: bluestore hints = 1 Maintaining 16 concurrent writes of 4194304 bytes to objects of size 4194304 for up to 10 seconds or 0 objects Object prefix: benchmark_data_stor-vm1_8786 sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg lat(s) 0 0 0 0 0 0 - 0 1 16 63 47 187.989 188 0.620617 0.285428 2 16 134 118 235.976 284 0.195319 0.250789 3 16 209 193 257.306 300 0.198448 0.239798 4 16 282 266 265.972 292 0.232927 0.233386 5 16 362 346 276.771 320 0.222398 0.226373 6 16 429 413 275.303 268 0.193111 0.226703 7 16 490 474 270.828 244 0.0879974 0.228776 8 16 562 546 272.97 288 0.125843 0.230455 9 16 625 609 270.637 252 0.145847 0.232388 10 16 701 685 273.97 304 0.411055 0.230831 Total time run: 10.161789 Total writes made: 702 Write size: 4194304 Object size: 4194304 Bandwidth (MB/sec): 276.329 Stddev Bandwidth: 38.1925 Max bandwidth (MB/sec): 320 Min bandwidth (MB/sec): 188 Average IOPS: 69 Stddev IOPS: 9 Max IOPS: 80 Min IOPS: 47 Average Latency(s): 0.231391 Stddev Latency(s): 0.107305 Max latency(s): 0.774406 Min latency(s): 0.0828756 Cleaning up (deleting benchmark objects) Removed 702 objects Clean up completed and total clean up time :1.190687 Thanks, Dan
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com