Neither made a difference.  I also have a glusterFS cluster with two nodes
in replicating mode residing on 1TB drives:

[root@triton speed]# dd conv=fdatasync if=/dev/zero of=/mnt/speed/test.out
bs=512k count=10000
10000+0 records in
10000+0 records out
5242880000 bytes (5.2 GB) copied, 43.573 s, 120 MB/s

... and Ceph:

[root@triton temp]# dd conv=fdatasync if=/dev/zero of=/mnt/temp/test.out
bs=512k count=10000
10000+0 records in
10000+0 records out
5242880000 bytes (5.2 GB) copied, 366.911 s, 14.3 MB/s


On Mon, Apr 8, 2013 at 4:29 PM, Mark Nelson <mark.nel...@inktank.com> wrote:

> On 04/08/2013 04:12 PM, Ziemowit Pierzycki wrote:
>
>> There is one SSD in each node.  IPoIB performance is about 7 gbps
>> between each host.  CephFS is mounted via kernel client.  Ceph version
>> is ceph-0.56.3-1.  I have a 1GB journal on the same drive as the OSD but
>> on a seperate file system split via LVM.
>>
>> Here is output of another test with fdatasync:
>>
>> [root@triton temp]# dd conv=fdatasync if=/dev/zero of=/mnt/temp/test.out
>> bs=512k count=10000
>> 10000+0 records in
>> 10000+0 records out
>> 5242880000 bytes (5.2 GB) copied, 359.307 s, 14.6 MB/s
>> [root@triton temp]# dd if=/mnt/temp/test.out of=/dev/null bs=512k
>> count=10000
>> 10000+0 records in
>> 10000+0 records out
>> 5242880000 bytes (5.2 GB) copied, 14.0521 s, 373 MB/s
>>
>
> Definitely seems off!  How many SSDs are involved and how fast are they
> each?  The MTU idea might have merit, but I honestly don't know enough
> about how well IPoIB handles giant MTUs like that.  One thing I have
> noticed on other IPoIB setups is that TCP autotuning can cause a ton of
> problems.  You may want to try disabling it on all of the hosts involved:
>
> echo 0 | tee /proc/sys/net/ipv4/tcp_**moderate_rcvbuf
>
> If that doesn't work, maybe try setting MTU to 9000 or 1500 if possible.
>
> Mark
>
>
>
>
>> The network traffic appears to match the transfer speeds shown here too.
>>   Writing is very slow.
>>
>>
>> On Mon, Apr 8, 2013 at 3:04 PM, Mark Nelson <mark.nel...@inktank.com
>> <mailto:mark.nelson@inktank.**com <mark.nel...@inktank.com>>> wrote:
>>
>>     Hi,
>>
>>     How many drives?  Have you tested your IPoIB performance with iperf?
>>       Is this CephFS with the kernel client?  What version of Ceph?  How
>>     are your journals configured? etc.  It's tough to make any
>>     recommendations without knowing more about what you are doing.
>>
>>     Also, please use conv=fdatasync when doing buffered IO writes with dd.
>>
>>     Thanks,
>>     Mark
>>
>>
>>     On 04/08/2013 03:00 PM, Ziemowit Pierzycki wrote:
>>
>>         Hi,
>>
>>         The first test was writing 500 mb file and was clocked at 1.2
>>         GBps.  The
>>         second test was writing 5000 mb file at 17 MBps.  The third test
>> was
>>         reading the file at ~400 MBps.
>>
>>
>>         On Mon, Apr 8, 2013 at 2:56 PM, Gregory Farnum <g...@inktank.com
>>         <mailto:g...@inktank.com>
>>         <mailto:g...@inktank.com <mailto:g...@inktank.com>>> wrote:
>>
>>              More details, please. You ran the same test twice and
>>         performance went
>>              up from 17.5MB/s to 394MB/s? How many drives in each node,
>>         and of what
>>              kind?
>>              -Greg
>>              Software Engineer #42 @ http://inktank.com | http://ceph.com
>>
>>
>>              On Mon, Apr 8, 2013 at 12:38 PM, Ziemowit Pierzycki
>>              <ziemo...@pierzycki.com <mailto:ziemo...@pierzycki.com**>
>>         <mailto:ziemo...@pierzycki.com
>>
>>         <mailto:ziemo...@pierzycki.com**>__>> wrote:
>>               > Hi,
>>               >
>>               > I have a 3 node SSD-backed cluster connected over
>>         infiniband (16K
>>              MTU) and
>>               > here is the performance I am seeing:
>>               >
>>               > [root@triton temp]# !dd
>>               > dd if=/dev/zero of=/mnt/temp/test.out bs=512k count=1000
>>               > 1000+0 records in
>>               > 1000+0 records out
>>               > 524288000 bytes (524 MB) copied, 0.436249 s, 1.2 GB/s
>>               > [root@triton temp]# dd if=/dev/zero
>>         of=/mnt/temp/test.out bs=512k
>>               > count=10000
>>               > 10000+0 records in
>>               > 10000+0 records out
>>               > 5242880000 bytes (5.2 GB) copied, 299.077 s, 17.5 MB/s
>>               > [root@triton temp]# dd if=/mnt/temp/test.out
>>         of=/dev/null bs=512k
>>               > count=1000010000+0 records in
>>               > 10000+0 records out
>>               > 5242880000 bytes (5.2 GB) copied, 13.3015 s, 394 MB/s
>>               >
>>               > Does that look right?  How do I check this is not a
>> network
>>              problem, because
>>               > I remember seeing a kernel issue related to large MTU.
>>               >
>>               > ______________________________**___________________
>>
>>               > ceph-users mailing list
>>               > ceph-users@lists.ceph.com
>>         <mailto:ceph-us...@lists.ceph.**com <ceph-users@lists.ceph.com>>
>>         <mailto:ceph-us...@lists.ceph.**__com
>>         <mailto:ceph-us...@lists.ceph.**com <ceph-users@lists.ceph.com>>>
>>               >
>>         
>> http://lists.ceph.com/__**listinfo.cgi/ceph-users-ceph._**_com<http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com>
>>         
>> <http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>> >
>>
>>               >
>>
>>
>>
>>
>>         ______________________________**___________________
>>
>>         ceph-users mailing list
>>         ceph-users@lists.ceph.com 
>> <mailto:ceph-us...@lists.ceph.**com<ceph-users@lists.ceph.com>
>> >
>>         
>> http://lists.ceph.com/__**listinfo.cgi/ceph-users-ceph._**_com<http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com>
>>         
>> <http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>> >
>>
>>
>>     ______________________________**___________________
>>
>>     ceph-users mailing list
>>     ceph-users@lists.ceph.com 
>> <mailto:ceph-us...@lists.ceph.**com<ceph-users@lists.ceph.com>
>> >
>>     
>> http://lists.ceph.com/__**listinfo.cgi/ceph-users-ceph._**_com<http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com>
>>
>>     
>> <http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>> >
>>
>>
>>
>>
>> ______________________________**_________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>>
>>
> ______________________________**_________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to