Have you tried the multi-link? Did it helped? George.
On Apr 21, 2014, at 10:34 , Muhammad Ansar Javed <muhammad.an...@seecs.edu.pk> wrote: > I am able to achieve around 90% ( maximum 9390 Mbps) bandwidth on 10GE. There > were configuration issues disabling Intel Speedstep and Interrupt coalescing > helped in achieving expected network bandwidth. Varying send and recv buffer > sizes from 128 KB to 1 MB added just 50 Mbps with maximum bandwidth achieved > on 1 MB buffer size. > Thanks for support. > > > On Thu, Apr 17, 2014 at 6:05 AM, George Bosilca <bosi...@icl.utk.edu> wrote: > Muhammad, > > Our configuration of TCP is tailored for 1Gbs networks, so it’s performance > on 10G might be sub-optimal. That being said, the remaining of this email > will be speculation as I do not have access to a 10G system to test it. > > There are two things that I would test to see if I can improve the > performance. > > 1. The send and receive TCP suffer. These are handled by the btl_tcp_sndbuf > and btl_tcp_rcvbuf. By default these are set to 128K which is extremely small > for a 10G network. Try 256KB or maybe even 1M (you might need to fiddle with > your kernel to get here). > > 2. Add more links between the processes by increasing the default value for > btl_tcp_links to 2 or 4. > > You might also try to the following (but here I’m more skeptical). Try > pushing the value of btl_tcp_endpoint_cache up. This parameter is not to be > used eagerly in real applications with a complete communication pattern, but > for a benchmark it might be a good use. > > George. > > On Apr 16, 2014, at 06:30 , Muhammad Ansar Javed > <muhammad.an...@seecs.edu.pk> wrote: > >> Hi Ralph, >> Yes, you are right. I should have also tested NetPipe-MPI version earlier. I >> ran NetPipe-MPI version on 10G Ethernet and maximum bandwidth achieved is >> 5872 Mbps. Moreover, maximum bandwidth achieved by osu_bw test is 6080 Mbps. >> I have used OSU-Micro-Benchmarks version 4.3. >> >> >> On Wed, Apr 16, 2014 at 3:40 PM, Ralph Castain <r...@open-mpi.org> wrote: >> I apologize, but I am now confused. Let me see if I can translate: >> >> * you ran the non-MPI version of the NetPipe benchmark and got 9.5Gps on a >> 10Gps network >> >> * you ran iperf and got 9.61Gps - however, this has nothing to do with MPI. >> Just tests your TCP stack >> >> * you tested your bandwidth program on a 1Gps network and got about 90% >> efficiency. >> >> Is the above correct? If so, my actual suggestion was to run the MPI version >> of NetPipe and to use the OSB benchmark program as well. Your program might >> well be okay, but benchmarking is a hard thing to get right in a parallel >> world, so you might as well validate it by cross-checking the result. >> >> I suggest this mostly because your performance numbers are far worse than >> anything we've measured using those standard benchmarks, and so we should >> first ensure we aren't chasing a ghost. >> >> >> >> >> >> On Wed, Apr 16, 2014 at 1:41 AM, Muhammad Ansar Javed >> <muhammad.an...@seecs.edu.pk> wrote: >> Yes, I have tried NetPipe-Java and iperf for bandwidth and configuration >> test. NetPipe Java achieves maximum 9.40 Gbps while iperf achieves maximum >> 9.61 Gbps bandwidth. I have also tested my bandwidth program on 1Gbps >> Ethernet connection and it achieves 901 Mbps bandwidth. I am using the same >> program for 10G network benchmarks. Please find attached source file of >> bandwidth program. >> >> As far as --bind-to core is concerned, I think it is working fine. Here is >> output of --report-bindings switch. >> [host3:07134] MCW rank 0 bound to socket 0[core 0[hwt 0]]: [B/././.] >> [host4:10282] MCW rank 1 bound to socket 0[core 0[hwt 0]]: [B/././.] >> >> >> >> >> On Tue, Apr 15, 2014 at 8:39 PM, Ralph Castain <r...@open-mpi.org> wrote: >> Have you tried a typical benchmark (e.g., NetPipe or OMB) to ensure the >> problem isn't in your program? Outside of that, you might want to explicitly >> tell it to --bind-to core just to be sure it does so - it's supposed to do >> that by default, but might as well be sure. You can check by adding >> --report-binding to the cmd line. >> >> >> On Apr 14, 2014, at 11:10 PM, Muhammad Ansar Javed >> <muhammad.an...@seecs.edu.pk> wrote: >> >>> Hi, >>> I am trying to benchmark Open MPI performance on 10G Ethernet network >>> between two hosts. The performance numbers of benchmarks are less than >>> expected. The maximum bandwidth achieved by OMPI-C is 5678 Mbps and I was >>> expecting around 9000+ Mbps. Moreover latency is also quite higher than >>> expected, ranging from 37 to 59 us. Here is complete set of numbers. >>> >>> Latency >>> Open MPI C >>> Size Time (us) >>> 1 37.76 >>> 2 37.75 >>> 4 37.78 >>> 8 55.17 >>> 16 37.89 >>> 32 39.08 >>> 64 37.78 >>> 128 59.46 >>> 256 39.37 >>> 512 40.39 >>> 1024 47.18 >>> 2048 47.84 >>> >>> >>> Bandwidth >>> Open MPI C >>> Size (Bytes) Bandwidth (Mbps) >>> 2048 412.22 >>> 4096 539.59 >>> 8192 827.73 >>> 16384 1655.35 >>> 32768 3274.3 >>> 65536 1995.22 >>> 131072 3270.84 >>> 262144 4316.22 >>> 524288 5019.46 >>> 1048576 5236.17 >>> 2097152 5362.61 >>> 4194304 5495.2 >>> 8388608 5565.32 >>> 16777216 5678.32 >>> >>> >>> My environments consists of two hosts having point-to-point (switch-less) >>> 10Gbps Ethernet connection. Environment (OS, user, directory structure >>> etc) on both hosts is exactly same. There is no NAS or shared file system >>> between both hosts. Following are configuration and job launching commands >>> that I am using. Moreover, I have attached output of script ompi_info --all. >>> >>> Configuration commmand: ./configure --enable-mpi-java >>> --prefix=/home/mpj/installed/openmpi_installed CC=/usr/bin/gcc >>> --disable-mpi-fortran >>> >>> Job launching command: mpirun -np 2 -hostfile machines -npernode 1 >>> ./latency.out >>> >>> Are these numbers okay? If not then please suggest performance tuning >>> steps... >>> >>> Thanks >>> >>> -- >>> Ansar Javed >>> HPC Lab >>> SEECS NUST >>> Contact: +92 334 438 9394 >>> Email: muhammad.an...@seecs.edu.pk >>> <ompi_info.tar.bz2>_______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> >> -- >> Regards >> >> >> Ansar Javed >> HPC Lab >> SEECS NUST >> Contact: +92 334 438 9394 >> Email: muhammad.an...@seecs.edu.pk >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> >> -- >> Regards >> >> Ansar Javed >> HPC Lab >> SEECS NUST >> Contact: +92 334 438 9394 >> Email: muhammad.an...@seecs.edu.pk >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > -- > Regards > > Ansar Javed > HPC Lab > SEECS NUST > Contact: +92 334 438 9394 > Email: muhammad.an...@seecs.edu.pk > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users