I have the following setup:
Node1 = 8 x SSD
Node2 = 6 x SATA
Node3 = 6 x SATA
Client1
All Cisco UCS running RHEL6.5 + kernel 3.18.0 + ceph 0.88.

A "dd bs=4k oflag=direct" test directly on a OSD disk shows me:
Node1 = 60MB/s
Node2 = 30MB/s
Node2 = 30MB/s

I've created 2 pools, each size=1, pg_num=1024.
I've created a rbd image, formatted it ext4 (bs=4k), but also xfs.
A "dd bs=4k oflag=direct"  test on that image shows me   5 MB/s.
A "dd bs=4M oflag=direct"  test on that image shows me 150 MB/s.
A "dd bs=32M oflag=direct" test on that image shows me 260 MB/s.
A "rados bench write"      test on that pool  shows me 560 MB/s.

What am i doing wrong?
Why is a 4kb block size write so slow?

Thanks for any help...


Samuel Terburg




_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to