There is one SSD in each node. IPoIB performance is about 7 gbps between each host. CephFS is mounted via kernel client. Ceph version is ceph-0.56.3-1. I have a 1GB journal on the same drive as the OSD but on a seperate file system split via LVM.
Here is output of another test with fdatasync: [root@triton temp]# dd conv=fdatasync if=/dev/zero of=/mnt/temp/test.out bs=512k count=10000 10000+0 records in 10000+0 records out 5242880000 bytes (5.2 GB) copied, 359.307 s, 14.6 MB/s [root@triton temp]# dd if=/mnt/temp/test.out of=/dev/null bs=512k count=10000 10000+0 records in 10000+0 records out 5242880000 bytes (5.2 GB) copied, 14.0521 s, 373 MB/s The network traffic appears to match the transfer speeds shown here too. Writing is very slow. On Mon, Apr 8, 2013 at 3:04 PM, Mark Nelson <mark.nel...@inktank.com> wrote: > Hi, > > How many drives? Have you tested your IPoIB performance with iperf? Is > this CephFS with the kernel client? What version of Ceph? How are your > journals configured? etc. It's tough to make any recommendations without > knowing more about what you are doing. > > Also, please use conv=fdatasync when doing buffered IO writes with dd. > > Thanks, > Mark > > > On 04/08/2013 03:00 PM, Ziemowit Pierzycki wrote: > >> Hi, >> >> The first test was writing 500 mb file and was clocked at 1.2 GBps. The >> second test was writing 5000 mb file at 17 MBps. The third test was >> reading the file at ~400 MBps. >> >> >> On Mon, Apr 8, 2013 at 2:56 PM, Gregory Farnum <g...@inktank.com >> <mailto:g...@inktank.com>> wrote: >> >> More details, please. You ran the same test twice and performance went >> up from 17.5MB/s to 394MB/s? How many drives in each node, and of what >> kind? >> -Greg >> Software Engineer #42 @ http://inktank.com | http://ceph.com >> >> >> On Mon, Apr 8, 2013 at 12:38 PM, Ziemowit Pierzycki >> <ziemo...@pierzycki.com <mailto:ziemo...@pierzycki.com**>> wrote: >> > Hi, >> > >> > I have a 3 node SSD-backed cluster connected over infiniband (16K >> MTU) and >> > here is the performance I am seeing: >> > >> > [root@triton temp]# !dd >> > dd if=/dev/zero of=/mnt/temp/test.out bs=512k count=1000 >> > 1000+0 records in >> > 1000+0 records out >> > 524288000 bytes (524 MB) copied, 0.436249 s, 1.2 GB/s >> > [root@triton temp]# dd if=/dev/zero of=/mnt/temp/test.out bs=512k >> > count=10000 >> > 10000+0 records in >> > 10000+0 records out >> > 5242880000 bytes (5.2 GB) copied, 299.077 s, 17.5 MB/s >> > [root@triton temp]# dd if=/mnt/temp/test.out of=/dev/null bs=512k >> > count=1000010000+0 records in >> > 10000+0 records out >> > 5242880000 bytes (5.2 GB) copied, 13.3015 s, 394 MB/s >> > >> > Does that look right? How do I check this is not a network >> problem, because >> > I remember seeing a kernel issue related to large MTU. >> > >> > ______________________________**_________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> <mailto:ceph-us...@lists.ceph.**com<ceph-users@lists.ceph.com> >> > >> > >> http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> >> >> > >> >> >> >> >> ______________________________**_________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> >> >> > ______________________________**_________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com