Noam, Gus and List,

Did you statically link your openmpi when you built it?  If you did (the
default is NOT to do this) then that could explain the discrepancy.

Jim 

> -----Original Message-----
> From: users-boun...@open-mpi.org 
> [mailto:users-boun...@open-mpi.org] On Behalf Of Noam Bernstein
> Sent: Wednesday, June 24, 2009 9:38 AM
> To: Open MPI Users
> Subject: Re: [OMPI users] 50% performance reduction due to 
> OpenMPI v 1.3.2forcing all MPI traffic over Ethernet instead 
> of using Infiniband
> 
> 
> On Jun 23, 2009, at 6:19 PM, Gus Correa wrote:
> 
> > Hi Jim, list
> >
> > On my OpenMPI 1.3.2 ompi_info -config gives:
> >
> > Wrapper extra LIBS: -lrdmacm -libverbs -ltorque -lnuma -ldl -Wl,-- 
> > export-dynamic -lnsl -lutil -lm -ldl
> >
> > Yours doesn't seem to have the IB libraries: -lrdmacm -libverbs
> >
> > So, I would guess your OpenMPI 1.3.2 build doesn't have IB support.
> 
> The second of these statements doesn't follow from the first.
> 
> My "ompi_info -config" returns
> 
> ompi_info -config | grep LIBS
>                Build LIBS: -lnsl -lutil  -lm
>        Wrapper extra LIBS:   -ldl   -Wl,--export-dynamic 
> -lnsl -lutil - 
> lm -ldl
> 
> But it does have openib
> 
> ompi_info | grep openib
>                   MCA btl: openib (MCA v2.0, API v2.0, 
> Component v1.3.2)
> 
> and osu_bibw returns
> 
> # OSU MPI Bi-Directional Bandwidth Test v3.0
> # Size     Bi-Bandwidth (MB/s)
> 4194304                1717.43
> 
> which it's sure not getting over ethernet.  I think Jeff 
> Squyres' test (ompi_info | grep openib) must be more definitive.
> 
>                                                               
>               Noam
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to