Hi Ralph,
Yes, you are right. I should have also tested NetPipe-MPI version earlier.
I ran NetPipe-MPI version on 10G Ethernet and maximum bandwidth achieved is
5872 Mbps. Moreover, maximum bandwidth achieved by osu_bw test is 6080
Mbps. I have used OSU-Micro-Benchmarks version 4.3.


On Wed, Apr 16, 2014 at 3:40 PM, Ralph Castain <r...@open-mpi.org> wrote:

> I apologize, but I am now confused. Let me see if I can translate:
>
> * you ran the non-MPI version of the NetPipe benchmark and got 9.5Gps on a
> 10Gps network
>
> * you ran iperf and got 9.61Gps - however, this has nothing to do with
> MPI. Just tests your TCP stack
>
> * you tested your bandwidth program on a 1Gps network and got about 90%
> efficiency.
>
> Is the above correct? If so, my actual suggestion was to run the MPI
> version of NetPipe and to use the OSB benchmark program as well. Your
> program might well be okay, but benchmarking is a hard thing to get right
> in a parallel world, so you might as well validate it by cross-checking the
> result.
>
> I suggest this mostly because your performance numbers are far worse than
> anything we've measured using those standard benchmarks, and so we should
> first ensure we aren't chasing a ghost.
>
>
>
>
>
> On Wed, Apr 16, 2014 at 1:41 AM, Muhammad Ansar Javed <
> muhammad.an...@seecs.edu.pk> wrote:
>
>> Yes, I have tried NetPipe-Java and iperf for bandwidth and configuration
>> test. NetPipe Java achieves maximum 9.40 Gbps while iperf achieves maximum
>> 9.61 Gbps bandwidth. I have also tested my bandwidth program on 1Gbps
>> Ethernet connection and it achieves 901 Mbps bandwidth. I am using the same
>> program for 10G network benchmarks. Please find attached source file of
>> bandwidth program.
>>
>> As far as --bind-to core is concerned, I think it is working fine. Here
>> is output of --report-bindings switch.
>> [host3:07134] MCW rank 0 bound to socket 0[core 0[hwt 0]]: [B/././.]
>> [host4:10282] MCW rank 1 bound to socket 0[core 0[hwt 0]]: [B/././.]
>>
>>
>>
>>
>> On Tue, Apr 15, 2014 at 8:39 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>
>>> Have you tried a typical benchmark (e.g., NetPipe or OMB) to ensure the
>>> problem isn't in your program? Outside of that, you might want to
>>> explicitly tell it to --bind-to core just to be sure it does so - it's
>>> supposed to do that by default, but might as well be sure. You can check by
>>> adding --report-binding to the cmd line.
>>>
>>>
>>> On Apr 14, 2014, at 11:10 PM, Muhammad Ansar Javed <
>>> muhammad.an...@seecs.edu.pk> wrote:
>>>
>>> Hi,
>>> I am trying to benchmark Open MPI performance on 10G Ethernet network
>>> between two hosts. The performance numbers of benchmarks are less than
>>> expected. The maximum bandwidth achieved by OMPI-C is 5678 Mbps and I was
>>> expecting around 9000+ Mbps. Moreover latency is also quite higher than
>>> expected, ranging from 37 to 59 us. Here is complete set of numbers.
>>>
>>>
>>>
>>> *LatencyOpen MPI C    Size    Time (us)*
>>> 1         37.76
>>> 2         37.75
>>> 4         37.78
>>> 8         55.17
>>> 16       37.89
>>> 32       39.08
>>> 64       37.78
>>> 128     59.46
>>> 256     39.37
>>> 512     40.39
>>> 1024   47.18
>>> 2048   47.84
>>>
>>>
>>>
>>>
>>> *BandwidthOpen MPI C    Size (Bytes)    Bandwidth (Mbps)*
>>> 2048               412.22
>>> 4096               539.59
>>> 8192               827.73
>>> 16384             1655.35
>>> 32768             3274.3
>>> 65536             1995.22
>>> 131072           3270.84
>>> 262144           4316.22
>>> 524288           5019.46
>>> 1048576         5236.17
>>> 2097152         5362.61
>>> 4194304         5495.2
>>> 8388608         5565.32
>>> 16777216       5678.32
>>>
>>>
>>> My environments consists of two hosts having point-to-point
>>> (switch-less) 10Gbps Ethernet connection.  Environment (OS, user, directory
>>> structure etc) on both hosts is exactly same. There is no NAS or shared
>>> file system between both hosts. Following are configuration and job
>>> launching commands that I am using. Moreover, I have attached output of
>>> script ompi_info --all.
>>>
>>> Configuration commmand: ./configure --enable-mpi-java
>>> --prefix=/home/mpj/installed/openmpi_installed CC=/usr/bin/gcc
>>> --disable-mpi-fortran
>>>
>>> Job launching command: mpirun -np 2 -hostfile machines -npernode 1
>>> ./latency.out
>>>
>>> Are these numbers okay? If not then please suggest performance tuning
>>> steps...
>>>
>>> Thanks
>>>
>>> --
>>> Ansar Javed
>>> HPC Lab
>>> SEECS NUST
>>> Contact: +92 334 438 9394
>>> Email: muhammad.an...@seecs.edu.pk
>>>  <ompi_info.tar.bz2>_______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>
>>
>>
>> --
>> Regards
>>
>>
>> Ansar Javed
>> HPC Lab
>> SEECS NUST
>> Contact: +92 334 438 9394
>> Email: muhammad.an...@seecs.edu.pk
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>



-- 
Regards

Ansar Javed
HPC Lab
SEECS NUST
Contact: +92 334 438 9394
Email: muhammad.an...@seecs.edu.pk

Reply via email to