A few suggestions: - Try using Open MPI 1.8.1. It's the newest release, and has many improvements since the 1.6.x series.
- Try using "--mca btl openib,sm,self" (in both v1.6.x and v1.8.x). This allows Open MPI to use shared memory to communicate between processes on the same server, which can be a significant performance improvement over TCP or even IB. On Apr 23, 2014, at 11:10 AM, "Sasso, John (GE Power & Water, Non-GE)" <john1.sa...@ge.com> wrote: > I am running IMB (Intel MPI Benchmarks), the MPI-1 benchmarks, which was > built with Intel 12.1 compiler suite and OpenMPI 1.6.5 (and running w/ OMPI > 1.6.5). I decided to use the following for the mca parameters: > > --mca btl openib,tcp,self --mca btl_openib_receive_queues > X,9216,256,128,32:X,65536,256,128,32 > > where before, I always used “--mca btl openib,tcp,self”. This is for > performance analysis. On the SendRecv benchmark at 32 processes, IMB hangs. > I then tried: > > --mca btl_openib_receive_queues > X,128,256,192,128:X,2048,256,128,32:X,12288,256,128,32:X,65536,256,128,32 > > and IMB also hangs on the SendRecv benchmark, though at 64 processes. > > No errors have been recorded, not even in any system log files but ‘top’ > shows the MPI tasks running. How can I go about troubleshooting this hang, > as well as figuring out what (If any) MCA XRC-related parameters in > btl_openib_receive_queues I have to specify to get IMB running properly? I > did verify the IB cards are ConnectX. > > --john > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/